How to use an Encoder-Decoder LSTM to Echo Sequences of Random Integers

Last Updated on August 27, 2020

A powerful feature of Long Short-Term Memory (LSTM) recurrent neural networks is that they can remember observations over long sequence intervals.

This can be demonstrated by contriving a simple sequence echo problem where the entire input sequence or partial contiguous blocks of the input sequence are echoed as an output sequence.

Developing LSTM recurrent neural networks to address the sequence echo problem is both a good demonstration of the power of LSTMs and can be used to demonstrate state-of-the-art recurrent neural network architectures.

In this post, you will discover how to develop LSTMs to solve the full and partial sequence echo problem in Python using the Keras deep learning library.

After completing this tutorial, you will know:

  • How to generate random sequences of integers, represent them using a one hot encoding and frame the sequence as a supervised learning problem with input and output pairs.
  • How to develop a sequence-to-sequence LSTM to echo the entire input sequence as an output.
  • How to develop an encoder-decoder LSTM to echo partial sequences with lengths that differ from the input sequence.

Kick-start your project with my new book Long Short-Term Memory Networks With Python, including step-by-step tutorials and the Python source code files for all examples.

Let’s get started.

  • Update Jan/2020: Updated API for Keras 2.3 and TensorFlow 2.0.
How to use an Encoder-Decoder LSTM to Echo Sequences of Random Integers

How to use an Encoder-Decoder LSTM to Echo Sequences of Random Integers
Photo by hammy24601, some rights reserved.


This tutorial is divided into 3 parts; they are:

  1. Sequence Echo Problem
  2. Echo Whole Sequence (sequence-to-sequence model)
  3. Echo Partial Sequence (encoder-decoder model)


This tutorial assumes you have a Python SciPy environment installed. You can use either Python 2 or 3 with this example.

This tutorial assumes you have Keras v2.0 or higher installed with either the TensorFlow or Theano backend.

This tutorial also assumes you have scikit-learn, Pandas, NumPy, and Matplotlib installed.

If you need help setting up your Python environment, see this post:

Need help with LSTMs for Sequence Prediction?

Take my free 7-day email course and discover 6 different LSTM architectures (with code).

Click to sign-up and also get a free PDF Ebook version of the course.

Sequence Echo Problem

The echo sequence problem involves exposing an LSTM to a sequence of observations, one at a time, then asking the network to echo back a partial or full list of contiguous observations observed.

This forces the network to remember blocks of contiguous observations and is a great demonstration of the learning power of LSTM recurrent neural networks.

The first step is to write some code to generate a random sequence of integers and encode them for the network.

This involves 3 steps:

  1. Generate Random Sequence
  2. One Hot Encode Random Sequence
  3. Frame Encoded Sequences for Learning

Generate Random Sequence

We can generate random integers in Python using the randint() function that takes two parameters, indicating the range of integers from which to draw values.

In this tutorial, we will define the problem as having integer values between 0 and 99 with 100 unique values.

We can put this in a function called generate_sequence() that will generate a sequence of random integers of the desired length, with the default length set to 25 elements.

This function is listed below.

One Hot Encode Random Sequence

Once we have generated sequences of random integers, we need to transform them into a format that is suitable for training an LSTM network.

One option would be to rescale the integer to the range [0,1]. This would work and would require that the problem be phrased as regression.

I am interested in predicting the right number, not a number close to the expected value. This means I would prefer to frame the problem as classification rather than regression, where the expected output is a class and there are 100 possible class values.

In this case, we can use a one hot encoding of the integer values where each value is represented by a 100 elements binary vector that is all “0” values except the index of the integer, which is marked 1.

The function below called one_hot_encode() defines how to iterate over a sequence of integers and create a binary vector representation for each and returns the result as a 2-dimensional array.

We also need to decode the encoded values so that we can make use of the predictions, in this case, just review them.

The one hot encoding can be inverted by using the argmax() NumPy function that returns the index of the value in the vector with the largest value.

The function below, named one_hot_decode(), will decode an encoded sequence and can be used to later decode predictions from our network.

Frame Encoded Sequences for Learning

Once the sequences have been generated and encoded, they must be organized into a framing suitable for learning.

This involves organizing the linear sequence into input (X) and output (y) pairs.

For example, the sequence [1, 2, 3, 4, 5] can be framed as a sequence prediction problem with 2 inputs (t and t-1) and 1 output (t-1) as follows:

Note that the absence of data is marked with a NaN value. These rows can be padded with special characters and masked. Alternatively, and more simply, these rows can be removed from the dataset, at the cost of providing fewer examples from a sequence from which to learn. This latter approach will be the approach used in this example.

We will use the Pandas shift() function to create shifted versions of the encoded sequence. The rows with missing data are removed with the dropna() function.

We can then specify the input and output data from the shifted data frame. The data must be 3-dimensional for use with sequence prediction LSTMs. Both input and output require the dimensions [samples, timesteps, features], where samples is the number of rows, timesteps is the number of lag observations from which to learn to make predictions, and features is the number of separate values (e.g. 100 for the one hot encoding).

The code below defines the to_supervised() function that implements this functionality. It takes two parameters to specify the number of encoded integers to use as input and output. The number of output integers must be less than or equal to the number of inputs and is counted from the oldest observation.

For example, if 5 and 1 were specified for input and output on the sequence [1, 2, 3, 4, 5], then the function would return a single row in X with [1, 2, 3, 4, 5] and a single row in y with [1].

For more on converting sequences to supervised learning problems, see the posts:

Complete Example

We can tie all of this together.

Below is the complete code listing for generating a sequence of 25 random integers and encoding each integer as a binary vector, separating them into X,y pairs for learning, then printing the decoded pairs for review.

Note: Your results may vary given the stochastic nature of the algorithm or evaluation procedure, or differences in numerical precision. Consider running the example a few times and compare the average outcome.

The sequence was split into input sequences of 5 numbers and output sequences of the 3 oldest observations from the input sequences.

Now that we know how to prepare and represent random sequences of integers, we can look at using LSTMs to learn them.

Echo Whole Sequence
(sequence-to-sequence model)

In this section, we will develop an LSTM for a simple framing of the problem, that is the prediction or reproduction of the entire input sequence.

That is, given a fixed input sequence, such as 5 random integers, output the same sequence. This may not sound like a simple framing of the problem, but it is because the network architecture required to address it is straightforward.

We will generate random sequences of 25 integers and frame them as input-output pairs of 5 values. We will create a convenience function named get_data() that we will use to create encoded X,y pairs of random integers using all of the functionality prepared in the previous section. This function is listed below.

This function will be called with the parameters 5 and 5 to create 21 samples with 5 timesteps of 100 features as inputs and the same as outputs (21 rather than 25 because some rows with missing values due to the shifting of the sequence are removed).

We can now develop an LSTM for this problem. We will use a stateful LSTM and explicitly reset the internal state at the end of training for each generated sample. Maintaining internal state within the network across the samples in a sequence may or may not be required as the context required for learning will be provided as timesteps; nevertheless, this additional state may be helpful.

Let’s start off by defining the expected dimensions of the input data as 5 timesteps of 100 features. Because we are using a stateful LSTM, this will be specified using the batch_input_shape argument instead of the input_shape. The LSTM hidden layer will use 20 memory units, which should be more than sufficient to learn this problem.

A batch size of 7 will be used. The batch size must be a factor of the number of training samples (in this case 21) and defines the number samples after which the weights in the LSTM are updated. This means weights will be updated 3 times for each random sequence the network is trained on.

We want the output layer to output one integer at a time, one for each input observed.

We will define the output layer as a fully connected layer (Dense) with 100 neurons for each of the 100 possible integer values in the one hot encoding. Because we are using a one hot encoding and framing the problem as multi-class classification, we can use the softmax activation function in the Dense layer.

We need to wrap this output layer in a TimeDistributed layer. This is to ensure that we can use the output layer to predict one integer at a time for each of the items in the input sequence. This is key so that we are implementing a true many-to-many model (e.g. sequence-to-sequence), rather than a many-to-one model where a one-shot vector output is created based on the internal state and value of the last observation in the input sequence (e.g. the output layer outputting 5*100 values at once).

This requires that the previous LSTM layer returns sequences by setting return_sequences=True (e.g. one output for each observation in the input sequence rather than one output for the entire input sequence). The TimeDistributed layer performs the trick of applying each slice of the sequence from the LSTM layer as inputs to the wrapped Dense layer so that one integer can be predicted at a time.

We will use the log loss function suitable for multi-class classification problems (categorical_crossentropy) and the efficient ADAM optimization algorithm with default hyperparameters.

In addition to reporting the log loss each epoch, we will also report the classification accuracy to get an idea of how our model is training.

To ensure the network does not memorize the problem, and the network learns to generalize a solution for all possible input sequences, we will generate a new random sequence each training epoch. Internal state within the LSTM hidden layer will reset at the end of each epoch. We will fit the model for 500 training epochs.

Once fit, we will evaluate the model by making a prediction on one new random sequence of integers and compare the decoded expected output sequences to the predicted sequences.

Putting this all together, the complete code listing is provided below.

Running the example prints the log loss and accuracy each epoch. The run ends by generating a new random sequence and comparing expected sequences to predicted sequences.

Note: Your results may vary given the stochastic nature of the algorithm or evaluation procedure, or differences in numerical precision. Consider running the example a few times and compare the average outcome.

Given the chosen configuration, the network will converge to 100% accuracy pretty much every time it is run.

Echo Partial Sequence
(encoder-decoder model)

So far, so good, but what if we want the length of the output sequence to differ from the length of the input sequence?

That is, we want to echo the first 2 observations from the input sequence of 5 observations:

This is still a sequence-to-sequence prediction problem, but requires a change to the network architecture.

One way would be to have the output sequence the same length, and use padding to fill out the output sequence to the same length.

Alternately we can use a more elegant solution. We can implement an encoder-decoder network that allows a variable length output where an encoder learns an internal representation for the input sequence and the decoder reads the internal representation and learns how to create the output sequence of the same or differing length.

This is a more challenging problem for the network to solve, and in turn requires additional capacity (more memory units) and longer training (more epochs).

The input, first hidden LSTM layer, and TimeDistributed Dense output layer of the network stay the same, except we will increase the number of memory units from 20 to 150. We will also increase the batch size from 7 to 21 so that weight updates are performed at the end of all samples of a random sequence. This was found to result in faster learning for this network after some experimentation.

The first hidden layer is the encoder.

We must add an additional hidden LSTM layer to act as the decoder. Again, we will use 150 memory units in this layer, and as in the previous example, the layer before the TimeDistributed layer will return sequences instead of a single value.

These two layers do not fit neatly together. The encoder layer will output a 2D array (21, 150) and the decoder expects a 3D array as input (21, ?, 150).

We address this problem by adding a RepeatVector() layer between the encoder and decoder and ensure that the output of the encoder is repeated a suitable number of times to match the length of the output sequence. In this case, 2 times for the two timesteps in the output sequence.

The LSTM network is therefore defined as:

The number of training epochs was increased from 500 to 5,000 to account for the additional capacity of the network.

The rest of the example is the same.

Tying this together, the complete code listing is provided below.

Running the example will display log loss and accuracy on randomly generated sequences each training epoch.

Note: Your results may vary given the stochastic nature of the algorithm or evaluation procedure, or differences in numerical precision. Consider running the example a few times and compare the average outcome.

The chosen configuration will mean that the model will converge to 100% classification accuracy.

A final randomly generated sequence is generated and expected vs predicted sequences are compared.


This section lists some possible extensions to the tutorial you may wish to explore.

  • Linear Representation. A categorical (one hot encoding) framing of the problem was used that dramatically increased the number of weights required to model this problem (100 per random integer). Explore using a linear representation (scale integers to values between 0-1) and model the problem as regression. See how this affects the skill of the system, required network size (memory units), and training time (epochs).
  • Mask Missing Values. When the sequence data was framed as a supervised learning problem, the rows with missing data were removed. Explore using a Masking layer or special values (e.g. -1) to allow the network to ignore or learn to ignore these values.
  • Echo Longer Sequences. The partial subsequence that was learned to be echoed was only 2 items long. Explore echoing longer sequences using the encoder-decoder network. Note, you will likely need larger hidden layers (more memory units) trained for longer (more epochs).
  • Ignore State. Care was taken to only clear state at the end of all samples of each random integer sequence and not shuffle samples within a sequence. This may not be needed. Explore and contrast model performance using a stateless LSTM with a batch size of 1 (weight updates and state reset after each sample of each sequence). I expect little to no change.
  • Alternate Network Topologies. Care was taken to use a TimeDistributed wrapper for the output layer to ensure that a many-to-many network was used to model the problem. Explore the sequence echo problem with a one-to-one configuration (timesteps taken as features) or a many-to-one (output taken as a vector of features) and see how this affects the required network size (memory units) and training time (epochs). I would expect it to require a larger network and take longer.

Did you explore any of these extensions?
Share your findings in the comments below.


In this tutorial, you discovered how to develop an LSTM recurrent neural network to echo sequences and partial sequences from randomly generated lists of integers.

Specifically, you learned:

  • How to generate random sequences of integers, represent them using a one hot encoding, and frame the problem as a supervised learning problem.
  • How to develop a sequence-to-sequence based LSTM network to echo an entire input sequence.
  • How to develop an encoder-decoder based LSTM network to echo partial input sequences with lengths that differ from the length of the input sequence.

Do you have any questions?
Post your questions in the comments and I will do my best to answer.

Develop LSTMs for Sequence Prediction Today!

Long Short-Term Memory Networks with Python

Develop Your Own LSTM models in Minutes

...with just a few lines of python code

Discover how in my new Ebook:
Long Short-Term Memory Networks with Python

It provides self-study tutorials on topics like:
CNN LSTMs, Encoder-Decoder LSTMs, generative models, data preparation, making predictions and much more...

Finally Bring LSTM Recurrent Neural Networks to
Your Sequence Predictions Projects

Skip the Academics. Just Results.

See What's Inside

41 Responses to How to use an Encoder-Decoder LSTM to Echo Sequences of Random Integers

  1. Avatar
    Ravi Annaswamy June 19, 2017 at 4:25 am #


    Thanks for yet another excellent walkthrough of a great idea in code using practical example.

    I remember seeing this mind-boggling power of the encoder decoder architecture especially when attention is added – in this great tutorial using Chainer framework. The
    system learns to reverse arbitrarily long strings!

    I would love to see this application using Keras.

    I am looking for your enc-dec-with-attn. tutorial.

    Also, You should compile your sequence processing tutorials into a book and I will be its first buyer 🙂


    • Avatar
      Jason Brownlee June 19, 2017 at 8:45 am #

      Thanks Ravi! I would like to write a 14-day crash course to LSTMs or something. I’m glad to hear you think it would be valuable.

      • Avatar
        Jack July 5, 2017 at 8:58 pm #

        That would be awesome Jason.

  2. Avatar
    Natallia Lundqvist June 28, 2017 at 12:41 am #

    Thank you for your effort to make DNNs more accessible. I follow your blog with great interest.
    New release of TensorFlow (API r.1.2) incorporates Keras to great extent. However, TimeDistributed layer has disappeared. I wonder if one should rewrite the code with new syntax, i.e., model = tf.contrib.keras.models.Sequential() instead of model = Sequential() ?? Another question, is there compatibility between tf.contrib.keras.layers.LSTM and tf.contrib.rnn.LSTMCell in the new TensorFlow??

    • Avatar
      Jason Brownlee June 28, 2017 at 6:27 am #

      It still exists in Keras:

      Sorry, I don’t know about the copy of Keras in the TF codebase.

      • Avatar
        Natallia Lundqvist July 3, 2017 at 8:20 pm #

        Yes, now it’s there, but “It looks like the import did not make it into the latest release, but is in master:” . I could proceed with:
        from tensorflow.contrib.keras.python.keras.layers.wrappers import TimeDistributed

        but I get another error message:

        File “C:\Users\natlun\AppData\Local\Continuum\Miniconda3\envs\tensorflow\lib\site-packages\tensorflow\contrib\keras\python\keras\layers\”, line 194, in call

        TypeError: rnn() got an unexpected keyword argument ‘input_length’

        However the code from your example works perfectly alright in older version of Keras + TensorFlow…

        • Avatar
          Jason Brownlee July 6, 2017 at 10:02 am #

          Consider using the standalone version of Keras not the TF contrib version.

  3. Avatar
    Natallia Lundqvist July 7, 2017 at 2:48 am #

    Yes, it’s the only way to make it work. I’v heard from Google team, that the issue is resolved on Master and would be implemented in the new release API r1.3 BUT no one knows when they’ll make the release.

    Another question. Why do you use your own one_hot encode while there is Keras version (actually, several ways to do one_hot_encode)??

  4. Avatar
    Mick July 8, 2017 at 1:15 am #

    I found that predicted result correct only first trying. Below are several outputs of lines starting from “# evaluate LSTM” represented in complete code listing. Input samples added too.

    Trying 1
    [21, 50, 39, 59, 63] Expected: [21, 50] Predicted: [21, 50]
    [50, 39, 59, 63, 3] Expected: [50, 39] Predicted: [50, 39]
    [39, 59, 63, 3, 90] Expected: [39, 59] Predicted: [39, 59]

    [67, 6, 10, 5, 75] Expected: [67, 6] Predicted: [67, 6]
    [6, 10, 5, 75, 59] Expected: [6, 10] Predicted: [6, 10]
    [10, 5, 75, 59, 95] Expected: [10, 5] Predicted: [10, 5]

    Trying 2
    [28, 35, 38, 30, 64] Expected: [28, 35] Predicted: [50, 50]
    [35, 38, 30, 64, 44] Expected: [35, 38] Predicted: [39, 39]
    [38, 30, 64, 44, 91] Expected: [38, 30] Predicted: [59, 59]

    [34, 42, 43, 70, 89] Expected: [34, 42] Predicted: [6, 6]
    [42, 43, 70, 89, 36] Expected: [42, 43] Predicted: [10, 10]
    [43, 70, 89, 36, 57] Expected: [43, 70] Predicted: [10, 5]

    Trying 3
    [77, 81, 11, 62, 88] Expected: [77, 81] Predicted: [50, 50]
    [81, 11, 62, 88, 53] Expected: [81, 11] Predicted: [39, 39]
    [11, 62, 88, 53, 82] Expected: [11, 62] Predicted: [59, 59]

    [68, 63, 62, 30, 44] Expected: [68, 63] Predicted: [6, 6]
    [63, 62, 30, 44, 52] Expected: [63, 62] Predicted: [10, 10]
    [62, 30, 44, 52, 75] Expected: [62, 30] Predicted: [5, 90]

    • Avatar
      Mick July 8, 2017 at 1:20 am #

      I used Jupyter for testing and put code to separate cell

    • Avatar
      Mick July 8, 2017 at 1:38 am #

      Yeah, all calling of “get_data” and “model.predict” will failure but first.

      • Avatar
        Michael February 23, 2018 at 3:50 am #

        Any way to fix this?

    • Avatar
      MyFly August 23, 2018 at 12:40 am #

      I ran into this problem also.. Is there any way to get repeatable rsults?

        • Avatar
          myfly August 24, 2018 at 1:03 am #

          It seems like predict can’t be run twice.. Thus – not able to get repeatble code..

          # evaluate LSTM
          X,y = get_data(n_in, n_out)
          yhat = model.predict(X, batch_size=batch_size, verbose=0)
          # decode all pairs
          for i in range(len(X)):
          print(‘Expected:’, one_hot_decode(y[i]), ‘Predicted’, one_hot_decode(yhat[i]))

          • Avatar
            Jason Brownlee August 24, 2018 at 6:12 am #

            Not quite, neural networks are stochastic models.

            Learn more here:

          • Avatar
            yaged July 18, 2023 at 5:14 pm #

            You have to reset the states of the model before each re-evalution like this:

            # evaluate LSTM
            model.reset_states() #This needs to be added before each re-evalution
            X,y = get_data(n_in, n_out)
            yhat = model.predict(X, batch_size=batch_size, verbose=0)
            # decode all pairs
            for i in range(len(X)):
            print(‘Expected:’, one_hot_decode(y[i]), ‘Predicted’, one_hot_decode(yhat[i]))

  5. Avatar
    Dr.Lafa August 2, 2017 at 4:29 am #

    Hello, how can i load sequences of text into X_train and Y_train?
    [“How”, “are”, “you”, “?”] in the X_train
    [“I’m”, “fine”] in Y_train

  6. Avatar
    Priyash October 25, 2017 at 5:37 pm #

    Hello ,

    I am new to the machine learning , in this article I am having difficulty understanding the fact :-

    Input layer 3d tensor of shape [7,5,100] it means it has 7 batches and every batch there are 5 words/timesteps in the sequence and each word can be represented in 100 dimensional vector up to know I got it right ?

    next part is ‘LSTM(20 , rest of the code)’ it means 20 lstm units will be in the hidden layer right ?

    My question is how the mapping is taking place between the 3d tensor shape [7,5,100] to 20 lstm nodes ? this part I don’t understand are we applying fully connection or dense or timedistributeddense internally in LSTM method ?

    • Avatar
      Jason Brownlee October 26, 2017 at 5:24 am #

      It would be 7 samples, 5 time steps, 100 features.

      Each node would receive 100 features each time step for each sample,

  7. Avatar
    Shane Strunk October 21, 2019 at 7:16 am #

    This Encoder/Decoder model seems infinitely simpler to implement than other Encoder/Decoder models i’ve seen. I’m guessing this simplicity is possible because this is a case where the size of the input is never going to change and can be specified in the input shape.

    • Avatar
      Jason Brownlee October 21, 2019 at 1:40 pm #

      The reason is because it is based on an autoencoder and appears to be just as effective.

  8. Avatar
    Syed November 21, 2019 at 1:01 am #

    Hi, Just a quick question. So when you train the decoder model, do you not need to prepend a start token at the beginning of the decoder inputs? If so, then how would you do that for integer sequences?

  9. Avatar
    Savath Saypadith February 13, 2020 at 9:31 pm #

    How can I create a custom loss when I use TimeDistributed?
    I get the error when I use a custom loss:

    InvalidArgumentError: 2 root error(s) found.
    (0) Invalid argument: You must feed a value for placeholder tensor ‘input_7’ with dtype float and shape [?,224,224,1]
    [[{{node input_7}}]]
    (1) Invalid argument: You must feed a value for placeholder tensor ‘input_7’ with dtype float and shape [?,224,224,1]
    [[{{node input_7}}]]
    0 successful operations.
    0 derived errors ignored.

  10. Avatar
    Baqar June 6, 2020 at 11:20 am #

    Hi Jason,

    Thanks for another interesting application using LSTM based AE. I was wondering if the same could be implemented using Attention with integers (No Translation).

    Can you please guide me how to do that or make another post?

    Thanks in advance.

    • Avatar
      Jason Brownlee June 6, 2020 at 1:27 pm #

      I don’t see why not.

      Thanks for the suggestion, I may be able to write about this in the future.

      • Avatar
        Baqar June 7, 2020 at 8:04 am #

        Thanks for the quick response. I am actually new to python at this stage and I actually tried to implement the same with stock price dataset. I keep on getting this error message as ‘list assignment index out of range’.

        Would you please mind shedding some light on it?


          • Avatar
            Baqar June 7, 2020 at 8:30 am #

            Thanks, but I used the data with proper indexing. The error is when I used the one_hot_encoding for an array of size approx 100k. Does, it have anything to do with the values, as most of them are stock prices e,g. 57500 (actual price $ * 10000)? As I have trouble understanding the for loop in one_hot_encoding function.

            I do apologise if this is a bit childish problem to you.

            I am a full-time subscriber now btw 🙂

          • Avatar
            Jason Brownlee June 7, 2020 at 1:12 pm #

            One hot encoding is for a categorical variable only.

            It sounds like you have a numerical variable in which case a one hot encoding would not be appropriate.

          • Avatar
            Baqar June 7, 2020 at 10:58 am #

            No issues, I have resolved the above now. However, I am stuck with the memory requirement issue as the input size of the lstm input to be around [9,70000].

            Is there any other way of encoding we can use here?


          • Avatar
            Baqar June 9, 2020 at 1:04 pm #

            Hi Jason,

            I have avoided one hot encoding as I am opting for an unsupervised model for integer variable. Does that mean in the above code, I can write X = y = input variable (1-D) ?

            And then how would the model work with input batch shape reduced to 2 dimensions instead of 3, more importantly how can I change it?

            Thanks again

          • Avatar
            Jason Brownlee June 9, 2020 at 1:22 pm #

            Not sure I follow. If it is unsupervised, perhaps you could use an autoencoder architecture instead:

          • Avatar
            Baqar June 9, 2020 at 2:33 pm #

            Thanks for your suggestion. I was initially asking about the addition of attention mechanism in the lstm auto-encoders using integer variables. But I got the gist now, I just have to try and find out a way to fit attention into it.

            If possible, I would be highly obliged if something like that could be explained by you.

  11. Avatar
    Gruhit Patel July 10, 2020 at 12:41 pm #

    Sir actually I’m working on generating random quotes based on labels applied to it. So how can I have a varying sequence output. Because output of my decoder is the same i applied to it and that is (batch_size, No_seq, 200) which is labels encoded to one hot vector.

    • Avatar
      Jason Brownlee July 10, 2020 at 1:53 pm #

      You can pre-define a long output and zero pad training examples.

      Or you can use a dynamic rnn model.

Leave a Reply