Last Updated on

Test datasets are small contrived datasets that let you test a machine learning algorithm or test harness.

The data from test datasets have well-defined properties, such as linearly or non-linearity, that allow you to explore specific algorithm behavior. The scikit-learn Python library provides a suite of functions for generating samples from configurable test problems for regression and classification.

In this tutorial, you will discover test problems and how to use them in Python with scikit-learn.

After completing this tutorial, you will know:

- How to generate multi-class classification prediction test problems.
- How to generate binary classification prediction test problems.
- How to generate linear regression prediction test problems.

Discover how to prepare data with pandas, fit and evaluate models with scikit-learn, and more in my new book, with 16 step-by-step tutorials, 3 projects, and full python code.

Let’s get started.

## Tutorial Overview

This tutorial is divided into 3 parts; they are:

- Test Datasets
- Classification Test Problems
- Regression Test Problems

## Test Datasets

A problem when developing and implementing machine learning algorithms is how do you know whether you have implemented them correctly. They seem to work even with bugs.

Test datasets are small contrived problems that allow you to test and debug your algorithms and test harness. They are also useful for better understanding the behavior of algorithms in response to changes in hyperparameters.

Below are some desirable properties of test datasets:

- They can be generated quickly and easily.
- They contain “known” or “understood” outcomes for comparison with predictions.
- They are stochastic, allowing random variations on the same problem each time they are generated.
- They are small and easily visualized in two dimensions.
- They can be scaled up trivially.

I recommend using test datasets when getting started with a new machine learning algorithm or when developing a new test harness.

scikit-learn is a Python library for machine learning that provides functions for generating a suite of test problems.

In this tutorial, we will look at some examples of generating test problems for classification and regression algorithms.

## Classification Test Problems

Classification is the problem of assigning labels to observations.

In this section, we will look at three classification problems: blobs, moons and circles.

### Blobs Classification Problem

The make_blobs() function can be used to generate blobs of points with a Gaussian distribution.

You can control how many blobs to generate and the number of samples to generate, as well as a host of other properties.

The problem is suitable for linear classification problems given the linearly separable nature of the blobs.

The example below generates a 2D dataset of samples with three blobs as a multi-class classification prediction problem. Each observation has two inputs and 0, 1, or 2 class values.

1 2 |
# generate 2d classification dataset X, y = make_blobs(n_samples=100, centers=3, n_features=2) |

The complete example is listed below.

1 2 3 4 5 6 7 8 9 10 11 12 13 |
from sklearn.datasets.samples_generator import make_blobs from matplotlib import pyplot from pandas import DataFrame # generate 2d classification dataset X, y = make_blobs(n_samples=100, centers=3, n_features=2) # scatter plot, dots colored by class value df = DataFrame(dict(x=X[:,0], y=X[:,1], label=y)) colors = {0:'red', 1:'blue', 2:'green'} fig, ax = pyplot.subplots() grouped = df.groupby('label') for key, group in grouped: group.plot(ax=ax, kind='scatter', x='x', y='y', label=key, color=colors[key]) pyplot.show() |

Running the example generates the inputs and outputs for the problem and then creates a handy 2D plot showing points for the different classes using different colors.

Note, your specific dataset and resulting plot will vary given the stochastic nature of the problem generator. This is a feature, not a bug.

We will use this same example structure for the following examples.

### Moons Classification Problem

The make_moons() function is for binary classification and will generate a swirl pattern, or two moons.

You can control how noisy the moon shapes are and the number of samples to generate.

This test problem is suitable for algorithms that are capable of learning nonlinear class boundaries.

The example below generates a moon dataset with moderate noise.

1 2 |
# generate 2d classification dataset X, y = make_moons(n_samples=100, noise=0.1) |

The complete example is listed below.

1 2 3 4 5 6 7 8 9 10 11 12 13 |
from sklearn.datasets import make_moons from matplotlib import pyplot from pandas import DataFrame # generate 2d classification dataset X, y = make_moons(n_samples=100, noise=0.1) # scatter plot, dots colored by class value df = DataFrame(dict(x=X[:,0], y=X[:,1], label=y)) colors = {0:'red', 1:'blue'} fig, ax = pyplot.subplots() grouped = df.groupby('label') for key, group in grouped: group.plot(ax=ax, kind='scatter', x='x', y='y', label=key, color=colors[key]) pyplot.show() |

Running the example generates and plots the dataset for review, again coloring samples by their assigned class.

### Circles Classification Problem

The make_circles() function generates a binary classification problem with datasets that fall into concentric circles.

Again, as with the moons test problem, you can control the amount of noise in the shapes.

This test problem is suitable for algorithms that can learn complex non-linear manifolds.

The example below generates a circles dataset with some noise.

1 2 |
# generate 2d classification dataset X, y = make_circles(n_samples=100, noise=0.05) |

The complete example is listed below.

1 2 3 4 5 6 7 8 9 10 11 12 13 |
from sklearn.datasets import make_circles from matplotlib import pyplot from pandas import DataFrame # generate 2d classification dataset X, y = make_circles(n_samples=100, noise=0.05) # scatter plot, dots colored by class value df = DataFrame(dict(x=X[:,0], y=X[:,1], label=y)) colors = {0:'red', 1:'blue'} fig, ax = pyplot.subplots() grouped = df.groupby('label') for key, group in grouped: group.plot(ax=ax, kind='scatter', x='x', y='y', label=key, color=colors[key]) pyplot.show() |

Running the example generates and plots the dataset for review.

## Regression Test Problems

Regression is the problem of predicting a quantity given an observation.

The make_regression() function will create a dataset with a linear relationship between inputs and the outputs.

You can configure the number of samples, number of input features, level of noise, and much more.

This dataset is suitable for algorithms that can learn a linear regression function.

The example below will generate 100 examples with one input feature and one output feature with modest noise.

1 2 |
# generate regression dataset X, y = make_regression(n_samples=100, n_features=1, noise=0.1) |

The complete example is listed below.

1 2 3 4 5 6 7 |
from sklearn.datasets import make_regression from matplotlib import pyplot # generate regression dataset X, y = make_regression(n_samples=100, n_features=1, noise=0.1) # plot regression dataset pyplot.scatter(X,y) pyplot.show() |

Running the example will generate the data and plot the X and y relationship, which, given that it is linear, is quite boring.

## Extensions

This section lists some ideas for extending the tutorial that you may wish to explore.

**Compare Algorithms**. Select a test problem and compare a suite of algorithms on the problem and report the performance.**Scale Up Problem**. Select a test problem and explore scaling it up, use progression methods to visualize the results, and perhaps explore model skill vs problem scale for a given algorithm.**Additional Problems**. The library provides a suite of additional test problems; write a code example for each to demonstrate how they work.

If you explore any of these extensions, I’d love to know.

## Further Reading

This section provides more resources on the topic if you are looking to go deeper.

## Summary

In this tutorial, you discovered test problems and how to use them in Python with scikit-learn.

Specifically, you learned:

- How to generate multi-class classification prediction test problems.
- How to generate binary classification prediction test problems.
- How to generate linear regression prediction test problems.

Do you have any questions?

Ask your questions in the comments below and I will do my best to answer.

Can the number of features for these datasets be greater than the examples given? For example, can the make_blobs function make datasets with 3+ features?

Try it.

Is there an analogous function in R?

There must be, I don’t know off hand sorry.

Hi,

I have built my model for gender prediction based on Text dataset using Multinomial Naive Bayes algorithm. However, I am trying to use my built model to make predictions on new real test dataset for Gender-based on Text. How do I achieve that? Can you please explain me the concept? Thanks.

This is a common question that I answer here:

https://machinelearningmastery.com/faq/single-faq/how-do-i-make-predictions

hi Jason , am working on credit card fraud detection where datasets are missing , can use that method to generate a datasets to validate my work , if no should abandon that work

best regard

If you do not have data, you cannot develop and test a model.

Alternately, if you have missing observations in a dataset, you have options:

https://machinelearningmastery.com/faq/single-faq/how-do-i-handle-missing-data

Why does make_blobs assign a classification y to the data points? Isn’t that the job of a classification algorithm?

Yes, but we need data to train the model.

Hey,

Have any idea on how to create a time series dataset using Brownian motion including trend and seasonality?

Sorry, I don’t have an example of Brownian motion.

How can I generate an imbalanced dataset? For example among 100 points I want 10 in one class and 90 in other class.

I’m sure the API can do it, but if not, generate with 100 examples in each class, then delete 90 examples from one class and 10 from the other.

can i generate a particular image detection by using this?

I don’t think so.

Hi Jason. Thanks for the great article. I already have a dataset that I want to increase its size. Maybe by copying some of the records but I’m looking for a more accurate way of doing it. Also do you know of a python library that can generate new data points out of a current dataset?

Many thanks,

Pamela

The simplest way is to copy records and add Gaussian noise with zero mean and a small stdev that makes sense for each dimension of your data.

Beyond that, you may want to look into resampling methods used by techniques such as SMOTE, etc.

Sorry, I don’t know of libraries that do this.

How would I plot something with more n_features? For example, in the blob generator, if I set n_features to 7, I get 7 columns of features. However, when I plot it, it only takes the first two columns as data for the plot. More importantly, the way it assigns a y-value seems to only be based on the first two feature columns as well – are the remaining features taken into account at all when it groups the data into specific clusters? I hope my question makes sense. Thank you.

You can choose the number of features and the number of features that contribute to the outcome.

Obviously, a 2D plot can only show two features at a time, you could create a matrix of each variable plotted against every other variable.

Thank you Jason, I confused the meaning of ‘centers’ with what normally would be equivalent to the y_train/y_test element (as the n_features element is basically the features in neural networks (X_train/X_test), so I falsely parallelized ‘centers’ with y_train/y_test in multivariate networks). Solves the graphing confusion as well.

Here, “center” referrs to an artificial cluster center for a samples that belong to a class.

i have to create a data.pkl and label.pkl files of some images with the dataset of some images .

how can i create a data and label.pkl form the data set of images ?

Perhaps load the data as numpy arrays and save the numpy arrays using the numpy save() function instead of using pickle?

Thank you, Jason, for this nice tutorial!

It helped me in finding a module in the sklearn by the name ‘datasets.make_regression’.

I am currently trying to understand how pca works and require to make some mock data of higher dimension than the feature itself.

In ‘datasets.make_regression’ the argument ‘n_feature’ is simple to understand, but ‘n_informative’ is confusing to me.

I desire my (initial) data to comprise of more feature columns than the actual ones and I try the following:

for, n_informative > n_feature, I get X.shape as (n,n_feature), where n is the total number of sample points.

Then, later on, I might want to carry out pca to reduce the dimension, which I seem to handle (say).

The question I want to ask is how do I obtain X.shape as (n, n_informative)?

The ‘n_informative’ argument controls how many of the input arguments are real or contribute to the outcome.

It sounds like you might want to set n_informative to the number of dimensions of your dataset.

Why n_features are used?

It specifies the number of variables we want in our problem, e.g. input variables.

hello there,

I have been asked to do a clustering using k Mean Algorithm for gene expression data and asked to provide the clustering result. Please provide me with the answer.

Sorry, I don’t have any tutorials on clustering at this stage.