Time Series Forecasting as Supervised Learning

Time series forecasting can be framed as a supervised learning problem.

This re-framing of your time series data allows you access to the suite of standard linear and nonlinear machine learning algorithms on your problem.

In this post, you will discover how you can re-frame your time series problem as a supervised learning problem for machine learning. After reading this post, you will know:

  • What supervised learning is and how it is the foundation for all predictive modeling machine learning algorithms.
  • The sliding window method for framing a time series dataset and how to use it.
  • How to use the sliding window for multivariate data and multi-step forecasting.

Let’s get started.

Time Series Forecasting as Supervised Learning

Time Series Forecasting as Supervised Learning
Photo by Jeroen Looyé, some rights reserved.

Supervised Machine Learning

The majority of practical machine learning uses supervised learning.

Supervised learning is where you have input variables (X) and an output variable (y) and you use an algorithm to learn the mapping function from the input to the output.

The goal is to approximate the real underlying mapping so well that when you have new input data (X), you can predict the output variables (y) for that data.

Below is a contrived example of a supervised learning dataset where each row is an observation comprised of one input variable (X) and one output variable to be predicted (y).

It is called supervised learning because the process of an algorithm learning from the training dataset can be thought of as a teacher supervising the learning process.

We know the correct answers; the algorithm iteratively makes predictions on the training data and is corrected by making updates. Learning stops when the algorithm achieves an acceptable level of performance.

Supervised learning problems can be further grouped into regression and classification problems.

  • Classification: A classification problem is when the output variable is a category, such as “red” and “blue” or “disease” and “no disease.”
  • Regression: A regression problem is when the output variable is a real value, such as “dollars” or “weight.” The contrived example above is a regression problem.

Stop learning Time Series Forecasting the slow way

Sign-up and get a FREE 7-day Time Series Forecasting Mini-Course

You will get:
...one lesson each day delivered to your inbox
...exclusive PDF ebook containing all lessons
...confidence and skills to work through your own projects

Download Your FREE Mini-Course

Sliding Window For Time Series Data

Time series data can be phrased as supervised learning.

Given a sequence of numbers for a time series dataset, we can restructure the data to look like a supervised learning problem. We can do this by using previous time steps as input variables and use the next time step as the output variable.

Let’s make this concrete with an example. Imagine we have a time series as follows:

We can restructure this time series dataset as a supervised learning problem by using the value at the previous time step to predict the value at the next time-step. Re-organizing the time series dataset this way, the data would look as follows:

Take a look at the above transformed dataset and compare it to the original time series. Here are some observations:

  • We can see that the previous time step is the input (X) and the next time step is the output (y) in our supervised learning problem.
  • We can see that the order between the observations is preserved, and must continue to be preserved when using this dataset to train a supervised model.
  • We can see that we have no previous value that we can use to predict the first value in the sequence. We will delete this row as we cannot use it.
  • We can also see that we do not have a known next value to predict for the last value in the sequence. We may want to delete this value while training our supervised model also.

The use of prior time steps to predict the next time step is called the sliding window method. For short, it may be called the window method in some literature. In statistics and time series analysis, this is called a lag or lag method.

The number of previous time steps is called the window width or size of the lag.

This sliding window is the basis for how we can turn any time series dataset into a supervised learning problem. From this simple example, we can notice a few things:

  • We can see how this can work to turn a time series into either a regression or a classification supervised learning problem for real-valued or labeled time series values.
  • We can see how once a time series dataset is prepared this way that any of the standard linear and nonlinear machine learning algorithms may be applied, as long as the order of the rows is preserved.
  • We can see how the width sliding window can be increased to include more previous time steps.
  • We can see how the sliding window approach can be used on a time series that has more than one value, or so-called multivariate time series.

We will explore some of these uses of the sliding window, starting next with using it to handle time series with more than one observation at each time step, called multivariate time series.

Sliding Window With Multivariate Time Series Data

The number of observations recorded for a given time in a time series dataset matters.

Traditionally, different names are used:

  • Univariate Time Series: These are datasets where only a single variable is observed at each time, such as temperature each hour. The example in the previous section is a univariate time series dataset.
  • Multivariate Time Series: These are datasets where two or more variables are observed at each time.

Most time series analysis methods, and even books on the topic, focus on univariate data. This is because it is the simplest to understand and work with. Multivariate data is often more difficult to work with. It is harder to model and often many of the classical methods do not perform well.

Multivariate time series analysis considers simultaneously multiple time series. … It is, in general, much more complicated than univariate time series analysis

— Page 1, Multivariate Time Series Analysis: With R and Financial Applications.

The sweet spot for using machine learning for time series is where classical methods fall down. This may be with complex univariate time series, and is more likely with multivariate time series given the additional complexity.

Below is another worked example to make the sliding window method concrete for multivariate time series.

Assume we have the contrived multivariate time series dataset below with two observations at each time step. Let’s also assume that we are only concerned with predicting measure2.

We can re-frame this time series dataset as a supervised learning problem with a window width of one.

This means that we will use the previous time step values of measure1 and measure2. We will also have available the next time step value for measure1. We will then predict the next time step value of measure2.

This will give us 3 input features and one output value to predict for each training pattern.

We can see that as in the univariate time series example above, we may need to remove the first and last rows in order to train our supervised learning model.

This example raises the question of what if we wanted to predict both measure1 and measure2 for the next time step?

The sliding window approach can also be used in this case.

Using the same time series dataset above, we can phrase it as a supervised learning problem where we predict both measure1 and measure2 with the same window width of one, as follows.

Not many supervised learning methods can handle the prediction of multiple output values without modification, but some methods, like artificial neural networks, have little trouble.

We can think of predicting more than one value as predicting a sequence. In this case, we were predicting two different output variables, but we may want to predict multiple time-steps ahead of one output variable.

This is called multi-step forecasting and is covered in the next section.

Sliding Window With Multi-Step Forecasting

The number of time steps ahead to be forecasted is important.

Again, it is traditional to use different names for the problem depending on the number of time-steps to forecast:

  • One-Step Forecast: This is where the next time step (t+1) is predicted.
  • Multi-Step Forecast: This is where two or more future time steps are to be predicted.

All of the examples we have looked at so far have been one-step forecasts.

There are are a number of ways to model multi-step forecasting as a supervised learning problem. We will cover some of these alternate ways in a future post.

For now, we are focusing on framing multi-step forecast using the sliding window method.

Consider the same univariate time series dataset from the first sliding window example above:

We can frame this time series as a two-step forecasting dataset for supervised learning with a window width of one, as follows:

We can see that the first row and the last two rows cannot be used to train a supervised model.

It is also a good example to show the burden on the input variables. Specifically, that a supervised model only has X1 to work with in order to predict both y1 and y2.

Careful thought and experimentation are needed on your problem to find a window width that results in acceptable model performance.

Further Reading

If you are looking for more resources on how to work with time series data as a machine learning problem, see the following two papers:


In this post, you discovered how you can re-frame your time series prediction problem as a supervised learning problem for use with machine learning methods.

Specifically, you learned:

  • Supervised learning is the most popular way of framing problems for machine learning as a collection of observations with inputs and outputs.
  • Sliding window is the way to restructure a time series dataset as a supervised learning problem.
  • Multivariate and multi-step forecasting time series can also be framed as supervised learning using the sliding window method.

Do you have any questions about the sliding window method or about this post?
Ask your questions in the comments below and I will do my best to answer.

Want to Develop Time Series Forecasts with Python?

Develop Your Own Forecasts in Minutes

...with just a few lines of python code

Discover how in my new Ebook:
Introduction to Time Series Forecasting With Python

It covers self-study tutorials and end-to-end projects on topics like:
Loading data, visualization, modeling, algorithm tuning, and much more...

Finally Bring Time Series Forecasting to
Your Own Projects

Skip the Academics. Just Results.

Click to learn more.


45 Responses to Time Series Forecasting as Supervised Learning

  1. Robert December 6, 2016 at 12:32 am #

    Thanks for the article. I understand the transformation. Now how do you separate the data into training and testing sets? Also, will the next article be working a simple example through to building a predictive model?

    • Jason Brownlee December 6, 2016 at 8:26 am #

      Great question Robert, I will have a post on this soon.

  2. Leo December 6, 2016 at 1:38 pm #

    Machine learning methods are not suitable for time series analysis. They do not take into account the relationship that exists between data values.

    • Jason Brownlee December 7, 2016 at 8:53 am #

      Interesting perspective Leo.

      Machine learning methods require this relationship is exposed to them explicitly in the form of a moving average, lag obs, seasonality indicators, etc. Just like linear regression does in ARIMA. No really a big leap here.

      Classical methods (like MA/AR/ARMA/ARIMA/and friends) breakdown when relationships are non-linear, obs are not iid, residuals are not gaussian, etc. Sometimes the complexity of the problem requires we try alternate methods.

      Finally, there are newer methods that can learn sequence, like LSTM recurrent neural networks. These methods have the potential to redefine an industry, just like has been done in speech recognition and computer vision.

  3. Leo December 7, 2016 at 12:12 pm #

    Machine learning methods require that there is no correlation between variables. This breaks down for time series where the lagged values are correlated.
    Moreover, there are many nonlinear time series methods like GARCH and its variants.

    • Jason Brownlee December 8, 2016 at 8:13 am #

      Great point, thanks Leo.

      The point about correlated inputs is true for many statistical methods, less true for others like trees, instance-based methods and even some neural nets (cnn and rnn).

      I think you’re spot on – most small univariate time series datasets will be satisfied with a classical statistical method. Perhaps LSTMs or decision trees on lagged vars can add something, perhaps not.

      When things get hairy in data with a time component (like movement prediction, gesture classification, …) perhaps ML is the way to go. I need to do a better job of fleshing out this detail.

      • yangsp March 17, 2017 at 8:03 pm #

        I tried it half a month ago, but it didn’t work well

    • dirk January 20, 2017 at 9:23 pm #

      Is that not a bit Bombastic.

      There are several quant hedge funds that have made and continue to make mind blowing returns through the use of ML methods and correlated variables in multivariate TS data.

      Maybe I’m missing something ?

  4. Leo December 8, 2016 at 10:38 am #

    Good point Jason. I guess I need to study LSTM.

  5. John January 8, 2017 at 4:47 pm #

    When do you public something about the Multi-Step Forecasting? 🙂

    • Jason Brownlee January 9, 2017 at 7:48 am #

      They are scheduled for later this month or early next month.

  6. Dehai January 20, 2017 at 9:19 am #

    The data generated from sensors of IoT or industrial machines are also typical time siries, and usually of huge amout, aka industrial big data.
    For this type of TS, many digital signal processing methods are used when being analysised, such as FFT, wavelet transform, euclidean distance.
    It seems that books discussing ML on TS usually don’t cover this DSP area. What do you think?

    • Jason Brownlee January 20, 2017 at 10:25 am #

      I agree Dehai.

      We can view these methods as data preparation/data transforms in the project process.

      Use of more advanced methods like FFT and wavelets requires knowledge of DSP which might be a step too far for devs looking to get into machine learning with little math background.

  7. Jay Urbain January 20, 2017 at 9:22 am #


    I had a project where I had to predict the likelihood of equipment failure from an event log. What worked pretty well was creating a training set from the event log with temporal target features that included whether or not a piece of equipment failed in the next 30, 60 days, etc. I also added temporal features for a piece of equipments past history, e.g., frequency of maintenance over different veriods, variance in measurements, etc. Could then apply any machine learning technique. Test set was created from last 20% of samples.
    — Jay Urbain

    • Jason Brownlee January 20, 2017 at 10:26 am #

      Very nice, thanks for sharing Jay!

    • xjackx February 13, 2017 at 10:25 am #

      Hi Jay,

      I am interested in finding out more about the predictive task you were involved with. Any chance you ahve a blog or can share more by email?

  8. Ziad January 20, 2017 at 1:32 pm #

    Jason, is using multi steps time lags with multivariate KNN or Random Forrest equivalent to transforming the feature space in similar way to kernel functions?

    I will also be curious to see how SVM can be used on multivariate problems.

    Thanks for the post.

    • Jason Brownlee January 21, 2017 at 10:23 am #

      I don’t think so Ziad, do you have a specific idea in mind?

  9. Kavitha Devi M K January 20, 2017 at 4:10 pm #

    In activity prediction application, the activity can be predicted only after multiple sequence of steps (multivariate time series data). Kindly suggest how to handle this problem for predicting the activity

    • Jason Brownlee January 21, 2017 at 10:24 am #

      Nice problem Kavitha. Sorry, I don’t have any examples of activity prediction. I don’t want to give you uninformed advice.

  10. pankaj January 20, 2017 at 8:42 pm #

    How would the time series restructuring be affected if we have 2 level or n level categorization within a time series. For example in case of sensor data we get it on each day and with-in the day say at every 5 seconds. The correlation may exist at the outer level i.e at day level but may not at internal level i.e at next sample (in seconds).
    Day1 Measure
    5PM 20
    5PM5Sc 22
    Day2 Measure
    5pm 25
    5pm5sc 27
    so on.

    • Jason Brownlee January 21, 2017 at 10:27 am #

      Great question pankaj.

      I would suggest resampling the data to a few different time scales and building a model on lag signals of each, then ensemble the predictions. Alsom build bigger models on lagged signals at each scale You want to give your models every opportunity to exploit the temporal structure in the problem.

  11. Okpako A. Ejaita January 21, 2017 at 6:10 pm #

    It was a great article. My question is not really on this topic.
    how can use capture the errors in a neural network for each instance of a data and print it out in java and now to interpolate on the captured errors so predict the errors.

  12. NGUYEN Quang Anh January 22, 2017 at 9:15 pm #

    This is great. Though the multi-step forecast is somewhat border me. If we make a data model with features, for example, 3 continuous lag, then it show that somehow, the next step would be build upon the value of these 3 data, like X(t) = a1.X(t-1) + a2.X(t-2) + a3.X(t-3). And what’s more, to predict further into the future, have we extended the width of the window ? In that case, as the number of features also extended, the size of training data also must be extended right ?

    • Jason Brownlee January 23, 2017 at 8:39 am #

      That is correct.

      There are two general approaches for a multi-step forecast: direct (one model for each future time step to be predicted) and recursive (use the one-step model again and again with predictions as inputs).

  13. Pranab January 23, 2017 at 3:05 pm #

    Nice article. You are proposing supervised learning for complex time series, instead of classical forecasting methods. Do you have any particular supervised learning method in mind? If so, what makes you think it will work better than NN based LSTM.

    You also mentioned, in response to a comment, that some ML techniques are not adversely impacted by correlated input. Can you please shed some light on your comment.

    • Jason Brownlee January 24, 2017 at 10:58 am #

      Hi Pranab,

      No specific method in mind, more of a methodology of framing time series forecasting as supervised learning, making it available to the suite of linear and nonlinear machine learning algorithms and ensemble methods. Not a new idea for sure.

      Sure, often decision trees are unflappable when it comes to irrelevant features and correlated features. In fact, often when there are unknown nonlinear interactions across features, accepting pairwise multicollinearity in input features results in better performing models.

  14. Hassine Saidane January 27, 2017 at 2:40 am #

    Hello Jason,

    This is a cery interesting. topic Have you considred forecasting one-step-ahead as a function of multi steps before. This will represent an output which is a function of several variables. The question of interest, by analogy to the traditionale mult-variate function, is how many variables (back step) to use and which ones are most significant to use through a variable selecion process.Variable selection could identify which time periods influence the analysis and forecat.

    This approach can greatly benefit the forecasting and anallysis of time series using all of machine learning algorithms.

    A colleague and I applied this approach. Four published papers on this work can be “googled using my name (Hassine Saidane)

    Happy continuation and thanks for sharing the article.

  15. sam February 8, 2017 at 11:51 am #

    Hi Jason,

    I am trying to predict customer attrition based on revenue trend as time series

    Month1 –> $ ; month2 –> $ as training data set.

    How can i use predictive algorithm to predict customer attrition based on the above training data ?


    • Jason Brownlee February 9, 2017 at 7:21 am #

      I would encourage you to re-read this post, it sells out exactly how to frame your problem Sam.

  16. Sam February 10, 2017 at 11:03 am #

    Thanks for your response Jason.I understood the above example.The above example seems to be predicting Y as regression value.But i am trying to predict Y as classification value (attrition = 1 or non attrition = 0)

    Example : Below is the time series of revenue where 1,2,3.. are the months and Y tell us if the customer attrited or not. Y will have only 2 values 1 or 0.

    So can i use the below format for my test data ?

    revenue1 revenue2 revenue3 …Y

    100 50 -25 1

    200 100 300 ….. 0

    Appreciate your help.


  17. Sam February 11, 2017 at 7:16 am #

    Thanks a ton Jason for your quick response.You made my day 🙂

  18. Anthony from Sydney February 22, 2017 at 3:22 pm #

    Dear Dr Jason,
    Two topics please
    (1) On cropping data and applying the model ‘to the real world’. I understand that cropping is done on the 0th and kth data points to get a 1:1 correspondence between data values at t and t-1. I assume from previous posts that you crop say the (k-10)th to kth data points, perform the successive 1 step ahead predictions and select the model based on the min(set of mse of all selected models) of the difference between the test and predicted models.
    (a)Is the idea to use the that model to predict the (k+1)th unknown.
    (b)Can we assume that the model you ‘trained’ will be acceptable when more data is acquired. In other words, what happens if you collect another x data points, and you want to predict the (k + x + 1) data point, can we assume that the model trained at k data points will work for the model at k + x data points? Or in other words, when do you ‘retrain’ the model.

    (2) On windowing the data: based on this blog, is the purpose of windowing the data to find the differences and train the differenced data to find the model. How can we make the assumption that the (k+1)th differenced observation can be predicted from the kth differenced observation.
    Thank you,
    Anthony from Sydney Australia

    • Jason Brownlee February 23, 2017 at 8:51 am #

      Hi Anthony,

      Sorry, I don’t understand what you mean by cropping. Perhaps you could give an example?

      Generally, we use all available historical data to make a one-step prediction (t+1) or a multi-step prediction (t+1, t+2, …, t+n). This applies when evaluating a model and when new data becomes available.

      Windowing is about framing a univariate time series into a supervised learning problem with lag obs as input features. This allows us to use traditional supervised learning algorithms to model the problem and make predictions.

      I hope that helps.

      • Anthony from Sydney February 23, 2017 at 10:02 am #

        Dear Dr Jason,
        I will rephrase both (1) and (2) into one.

        Perhaps I wasn’t very clear at all.

        Cropping. by cropping I mean remove the earliest, the 0th and the latest kth data points because there are no corresponding lagged values by virtue of lagging.
        data point value lagged data point array reference
        1 ? – this is cropped/pruned 0
        2 1 1
        3 2 2
        44 3 3
        5 4 4
        . .
        560 1234 k-1.
        ? X – this is cropped/pruned. k

        dataset available for processing
        datapoint lagged data point (array ref based on original data)
        2 1 1
        3 2 2
        44 3 3
        5 44 4
        . .
        560 1234 k-1

        This is the above dataset with the 0th and kth elements cropped/pruned from the original.
        I should have been clearer. I apologise.
        My questions
        (a) Based on the ‘new’ lagged dataset, how can you make a prediction for the (k + 1)th dataset given the kth data point is not available.In other words, are making a prediction for the (k+1)th data point based on the (k-1)th datapoint.

        (b) Perhaps I’m missing something, having read the other posts on ARIMA. How can we make the assumption that predicting the next data point is based on the previous data point when there may well be MA or AR or other kinds processes on the data? Or in other words how can we assume that differencingor windowing as in this tutorial/blog will be the basis of our training model?

        (c) Suppose you trained your model based on the original dataset. Suppose that as your system acquires more datapoints, won’t the original model that you trained become invalid. Say you got an extra 10 or 1000 datapoints, do you have to retrain your data because the coefficients of the original model may not be an adequate predictor for a larger dataset.

        Thank you again and I hope I have been clearer,
        Anthony of Sydney Australia

        • Jason Brownlee February 24, 2017 at 10:08 am #

          Hi Anthony,

          What is k? Is that a time step t? I think it is given context.

          If you want to forecast a new data point that is out of sample (t+1) beyond the training dataset, your model will use t-1, … t-n as inputs to make the forecast.

          This applied regardless of the type of model used. E.g. if you are using an AR, the inputs will be lagged obs. If MA, the inputs will be an autoregression of the lagged error series.

          If differencing is performed in the preparation of the model, it will have to be performed on any new data. The decision to difference or seasonally adjust is based on the data itself and your analysis of temporal structure like trends and seasonality.

          Yes, as new data comes in the model will need to be refit. This is not a requirement for all problems, but a good idea. To mimic this real world expectation, we evaluate models in the same way using walk-forward validation that does exactly this – refits a model each time a new ob is available and predicts the next out of sample ob.

          I hope this helps. I do cover all of this in my book, lesson by lesson.

  19. Anthony from Sydney February 23, 2017 at 10:06 am #

    Dear Dr Jason, apologies again, my original spaced data set example did not appear neat.
    In both the original and the cropped/pruned/windowed datasets, there are meant to be three columns consisting of the data, data lagged by 1, and the array index based on the original dataset.
    I don’t know how to get nicely spaced tabbed data when posting replies on this blog
    Anthony of Sydney

    • Jason Brownlee February 24, 2017 at 10:10 am #

      You can use the pre HTML tag, e.g.:

      • Anthony from Sydney February 24, 2017 at 11:36 am #

        On how to insert BBCode in forum replies

        * 1 ?
        * 2 1
        * 3 2
        * 4 3
        * ? 4
        This is an experiment in inserting HTML code on a forum reply.
        I hope this works,
        [b] Anthony [/b] [i] from Sydney [/i]

  20. Anthony of Sydney February 26, 2017 at 2:06 pm #

    Testing using the ‘pre’ enclosed in ”, inserting “this is a test message”, then ”

    Hope it works

  21. Nirikshith March 14, 2017 at 2:09 pm #

    Dear Jason,
    have you planned any blog on forecasting Multivariate Time Series? I went through your ARIMA post and it was good start point for me.
    #student #aspring data analyst

    • Jason Brownlee March 15, 2017 at 8:07 am #

      Thnaks Nirikshith.

      Yes, I hope to cover multivariate time series forecasting in depth soon.

Leave a Reply