# Text Generation With LSTM Recurrent Neural Networks in Python with Keras

Last Updated on September 3, 2020

Recurrent neural networks can also be used as generative models.

This means that in addition to being used for predictive models (making predictions) they can learn the sequences of a problem and then generate entirely new plausible sequences for the problem domain.

Generative models like this are useful not only to study how well a model has learned a problem, but to learn more about the problem domain itself.

In this post you will discover how to create a generative model for text, character-by-character using LSTM recurrent neural networks in Python with Keras.

After reading this post you will know:

• Where to download a free corpus of text that you can use to train text generative models.
• How to frame the problem of text sequences to a recurrent neural network generative model.
• How to develop an LSTM to generate plausible text sequences for a given problem.

Kick-start your project with my new book Deep Learning for Natural Language Processing, including step-by-step tutorials and the Python source code files for all examples.

Let’s get started.

Note: LSTM recurrent neural networks can be slow to train and it is highly recommend that you train them on GPU hardware. You can access GPU hardware in the cloud very cheaply using Amazon Web Services, see the tutorial here.

• Update Oct/2016: Fixed a few minor comment typos in the code.
• Update Mar/2017: Updated for Keras 2.0.2, TensorFlow 1.0.1 and Theano 0.9.0.
• Update Sep/2019: Updated for Keras 2.2.5 API.

Text Generation With LSTM Recurrent Neural Networks in Python with Keras
Photo by Russ Sanderlin, some rights reserved.

## Problem Description: Project Gutenberg

Many of the classical texts are no longer protected under copyright.

This means that you can download all of the text for these books for free and use them in experiments, like creating generative models. Perhaps the best place to get access to free books that are no longer protected by copyright is Project Gutenberg.

In this tutorial we are going to use a favorite book from childhood as the dataset: Alice’s Adventures in Wonderland by Lewis Carroll.

We are going to learn the dependencies between characters and the conditional probabilities of characters in sequences so that we can in turn generate wholly new and original sequences of characters.

This is a lot of fun and I recommend repeating these experiments with other books from Project Gutenberg, here is a list of the most popular books on the site.

These experiments are not limited to text, you can also experiment with other ASCII data, such as computer source code, marked up documents in LaTeX, HTML or Markdown and more.

You can download the complete text in ASCII format (Plain Text UTF-8) for this book for free and place it in your working directory with the filename wonderland.txt.

Now we need to prepare the dataset ready for modeling.

Project Gutenberg adds a standard header and footer to each book and this is not part of the original text. Open the file in a text editor and delete the header and footer.

The header is obvious and ends with the text:

The footer is all of the text after the line of text that says:

You should be left with a text file that has about 3,330 lines of text.

### Need help with LSTMs for Sequence Prediction?

Take my free 7-day email course and discover 6 different LSTM architectures (with code).

Click to sign-up and also get a free PDF Ebook version of the course.

## Develop a Small LSTM Recurrent Neural Network

In this section we will develop a simple LSTM network to learn sequences of characters from Alice in Wonderland. In the next section we will use this model to generate new sequences of characters.

Let’s start off by importing the classes and functions we intend to use to train our model.

Next, we need to load the ASCII text for the book into memory and convert all of the characters to lowercase to reduce the vocabulary that the network must learn.

Now that the book is loaded, we must prepare the data for modeling by the neural network. We cannot model the characters directly, instead we must convert the characters to integers.

We can do this easily by first creating a set of all of the distinct characters in the book, then creating a map of each character to a unique integer.

For example, the list of unique sorted lowercase characters in the book is as follows:

You can see that there may be some characters that we could remove to further clean up the dataset that will reduce the vocabulary and may improve the modeling process.

Now that the book has been loaded and the mapping prepared, we can summarize the dataset.

Running the code to this point produces the following output.

We can see that the book has just under 150,000 characters and that when converted to lowercase that there are only 47 distinct characters in the vocabulary for the network to learn. Much more than the 26 in the alphabet.

We now need to define the training data for the network. There is a lot of flexibility in how you choose to break up the text and expose it to the network during training.

In this tutorial we will split the book text up into subsequences with a fixed length of 100 characters, an arbitrary length. We could just as easily split the data up by sentences and pad the shorter sequences and truncate the longer ones.

Each training pattern of the network is comprised of 100 time steps of one character (X) followed by one character output (y). When creating these sequences, we slide this window along the whole book one character at a time, allowing each character a chance to be learned from the 100 characters that preceded it (except the first 100 characters of course).

For example, if the sequence length is 5 (for simplicity) then the first two training patterns would be as follows:

As we split up the book into these sequences, we convert the characters to integers using our lookup table we prepared earlier.

Running the code to this point shows us that when we split up the dataset into training data for the network to learn that we have just under 150,000 training pattens. This makes sense as excluding the first 100 characters, we have one training pattern to predict each of the remaining characters.

Now that we have prepared our training data we need to transform it so that it is suitable for use with Keras.

First we must transform the list of input sequences into the form [samples, time steps, features] expected by an LSTM network.

Next we need to rescale the integers to the range 0-to-1 to make the patterns easier to learn by the LSTM network that uses the sigmoid activation function by default.

Finally, we need to convert the output patterns (single characters converted to integers) into a one hot encoding. This is so that we can configure the network to predict the probability of each of the 47 different characters in the vocabulary (an easier representation) rather than trying to force it to predict precisely the next character. Each y value is converted into a sparse vector with a length of 47, full of zeros except with a 1 in the column for the letter (integer) that the pattern represents.

For example, when “n” (integer value 31) is one hot encoded it looks as follows:

We can implement these steps as below.

We can now define our LSTM model. Here we define a single hidden LSTM layer with 256 memory units. The network uses dropout with a probability of 20. The output layer is a Dense layer using the softmax activation function to output a probability prediction for each of the 47 characters between 0 and 1.

The problem is really a single character classification problem with 47 classes and as such is defined as optimizing the log loss (cross entropy), here using the ADAM optimization algorithm for speed.

There is no test dataset. We are modeling the entire training dataset to learn the probability of each character in a sequence.

We are not interested in the most accurate (classification accuracy) model of the training dataset. This would be a model that predicts each character in the training dataset perfectly. Instead we are interested in a generalization of the dataset that minimizes the chosen loss function. We are seeking a balance between generalization and overfitting but short of memorization.

The network is slow to train (about 300 seconds per epoch on an Nvidia K520 GPU). Because of the slowness and because of our optimization requirements, we will use model checkpointing to record all of the network weights to file each time an improvement in loss is observed at the end of the epoch. We will use the best set of weights (lowest loss) to instantiate our generative model in the next section.

We can now fit our model to the data. Here we use a modest number of 20 epochs and a large batch size of 128 patterns.

The full code listing is provided below for completeness.

Note: Your results may vary given the stochastic nature of the algorithm or evaluation procedure, or differences in numerical precision. Consider running the example a few times and compare the average outcome.

After running the example, you should have a number of weight checkpoint files in the local directory.

You can delete them all except the one with the smallest loss value. For example, when I ran this example, below was the checkpoint with the smallest loss that I achieved.

The network loss decreased almost every epoch and I expect the network could benefit from training for many more epochs.

In the next section we will look at using this model to generate new text sequences.

## Generating Text with an LSTM Network

Generating text using the trained LSTM network is relatively straightforward.

Firstly, we load the data and define the network in exactly the same way, except the network weights are loaded from a checkpoint file and the network does not need to be trained.

Also, when preparing the mapping of unique characters to integers, we must also create a reverse mapping that we can use to convert the integers back to characters so that we can understand the predictions.

Finally, we need to actually make predictions.

The simplest way to use the Keras LSTM model to make predictions is to first start off with a seed sequence as input, generate the next character then update the seed sequence to add the generated character on the end and trim off the first character. This process is repeated for as long as we want to predict new characters (e.g. a sequence of 1,000 characters in length).

We can pick a random input pattern as our seed sequence, then print generated characters as we generate them.

The full code example for generating text using the loaded LSTM model is listed below for completeness.

Running this example first outputs the selected random seed, then each character as it is generated.

Note: Your results may vary given the stochastic nature of the algorithm or evaluation procedure, or differences in numerical precision. Consider running the example a few times and compare the average outcome.

For example, below are the results from one run of this text generator. The random seed was:

The generated text with the random seed (cleaned up for presentation) was:

We can note some observations about the generated text.

• It generally conforms to the line format observed in the original text of less than 80 characters before a new line.
• The characters are separated into word-like groups and most groups are actual English words (e.g. “the”, “little” and “was”), but many do not (e.g. “lott”, “tiie” and “taede”).
• Some of the words in sequence make sense(e.g. “and the white rabbit“), but many do not (e.g. “wese tilel“).

The fact that this character based model of the book produces output like this is very impressive. It gives you a sense of the learning capabilities of LSTM networks.

The results are not perfect. In the next section we look at improving the quality of results by developing a much larger LSTM network.

## Larger LSTM Recurrent Neural Network

We got results, but not excellent results in the previous section. Now, we can try to improve the quality of the generated text by creating a much larger network.

We will keep the number of memory units the same at 256, but add a second layer.

We will also change the filename of the checkpointed weights so that we can tell the difference between weights for this network and the previous (by appending the word “bigger” in the filename).

Finally, we will increase the number of training epochs from 20 to 50 and decrease the batch size from 128 to 64 to give the network more of an opportunity to be updated and learn.

The full code listing is presented below for completeness.

Running this example takes some time, at least 700 seconds per epoch.

Note: Your results may vary given the stochastic nature of the algorithm or evaluation procedure, or differences in numerical precision. Consider running the example a few times and compare the average outcome.

After running this example you may achieved a loss of about 1.2. For example the best result I achieved from running this model was stored in a checkpoint file with the name:

Achieving a loss of 1.2219 at epoch 47.

As in the previous section, we can use this best model from the run to generate text.

The only change we need to make to the text generation script from the previous section is in the specification of the network topology and from which file to seed the network weights.

The full code listing is provided below for completeness.

One example of running this text generation script produces the output below.

Note: Your results may vary given the stochastic nature of the algorithm or evaluation procedure, or differences in numerical precision. Consider running the example a few times and compare the average outcome.

The randomly chosen seed text was:

The generated text with the seed (cleaned up for presentation) was :

We can see that generally there are fewer spelling mistakes and the text looks more realistic, but is still quite nonsensical.

For example the same phrases get repeated again and again like “said to herself” and “little“. Quotes are opened but not closed.

These are better results but there is still a lot of room for improvement.

## 10 Extension Ideas to Improve the Model

Below are 10 ideas that may further improve the model that you could experiment with are:

• Predict fewer than 1,000 characters as output for a given seed.
• Remove all punctuation from the source text, and therefore from the models’ vocabulary.
• Try a one hot encoded for the input sequences.
• Train the model on padded sentences rather than random sequences of characters.
• Increase the number of training epochs to 100 or many hundreds.
• Add dropout to the visible input layer and consider tuning the dropout percentage.
• Tune the batch size, try a batch size of 1 as a (very slow) baseline and larger sizes from there.
• Add more memory units to the layers and/or more layers.
• Experiment with scale factors (temperature) when interpreting the prediction probabilities.
• Change the LSTM layers to be “stateful” to maintain state across batches.

Did you try any of these extensions? Share your results in the comments.

## Resources

This character text model is a popular way for generating text using recurrent neural networks.

Below are some more resources and tutorials on the topic if you are interested in going deeper. Perhaps the most popular is the tutorial by Andrej Karpathy titled “The Unreasonable Effectiveness of Recurrent Neural Networks“.

## Summary

In this post you discovered how you can develop an LSTM recurrent neural network for text generation in Python with the Keras deep learning library.

After reading this post you know:

• How to train an LSTM network on text sequences and how to use the trained network to generate new sequences.
• How to develop stacked LSTM networks and lift the performance of the model.

## Develop Deep Learning models for Text Data Today!

#### Develop Your Own Text models in Minutes

...with just a few lines of python code

Discover how in my new Ebook:
Deep Learning for Natural Language Processing

It provides self-study tutorials on topics like:
Bag-of-Words, Word Embedding, Language Models, Caption Generation, Text Translation and much more...

### 426 Responses to Text Generation With LSTM Recurrent Neural Networks in Python with Keras

1. Avi Levy August 12, 2016 at 10:33 am #

Great post. Thanks!

• Jason Brownlee August 15, 2016 at 12:29 pm #

You’re welcome Avi.

• BARURI SAI AVINASH November 2, 2020 at 12:01 am #

Do the loss we get here is equal to number of bits per charecter ????

• Jason Brownlee November 2, 2020 at 6:40 am #

The loss represents the average difference between the expected and predicted probability distribution.

• Shreyas September 2, 2017 at 3:03 am #

Hi,
when i try to run the codes, I get an error with the weights file.
ValueError: Dimension 1 in both shapes must be equal, but are 52 and 44 for Assign_13 with input shapes [256,52], [256,44].
Can you please let me know what is happening

• Jason Brownlee September 2, 2017 at 6:15 am #

Confirm that you have copied and run all of the code and that your environment and libraries are all up to date.

• Shreyas Becker Lalitha Venkatramanan September 3, 2017 at 10:15 am #

Hi Jason,
Just updated all libraries. Still getting the same error.
Thanks!

• Jason Brownlee September 3, 2017 at 3:43 pm #

Sorry, it is not clear to me what your fault could be.

• Shreyas Becker Lalitha Venkatramanan September 3, 2017 at 8:21 pm #

Jason,
Thanks. I got it to work now!!! But I am getting random results will work on it further. But i feel like it is a good start. Thanks for the codes!
Thanks!

• Jason Brownlee September 4, 2017 at 4:30 am #

Well done on your progress, hang in there!

• Akash March 27, 2018 at 3:55 pm #

You have to make the shape equal, either [256,44], [256,44] (or) [256,52], [256,52].

• Senthil July 24, 2018 at 11:41 pm #

This was caused by the weights.hdf5 file being incompatible with the new data in the repository. I have updated the repo and it should work now.

• Ying May 27, 2020 at 1:50 pm #

I have the same issue. Where to download the hdf5 file. Thank youl

• oluwande adewoyin January 31, 2021 at 10:05 am #

pls how did you update the repo

2. Max August 16, 2016 at 3:30 pm #

I’m excited to try combining this with nltk… it’s probably going to be long and frustrating, but I’ll try to let you know my results.

Thanks for sharing!

• Jason Brownlee August 17, 2016 at 9:49 am #

Good luck Max, report back and let us know how you go.

• Al March 26, 2017 at 7:03 am #

Hi Max,

Did you get anywhere with combining NLTK with this?
Just curious…

3. Sviat September 10, 2016 at 10:23 pm #

I really like your article, thank you for sharing. I can launch your code but I have a crash after finalization of 1st epoch:
—————————————————
Traceback (most recent call last):
File “train.py”, line 61, in
model.fit(X, y, nb_epoch=20, batch_size=128, callbacks=callbacks_list)
File “/afs/in2p3.fr/home/s/sbilokin/.local/lib/python2.7/site-packages/keras/models.py”, line 620, in fit
sample_weight=sample_weight)
File “/afs/in2p3.fr/home/s/sbilokin/.local/lib/python2.7/site-packages/keras/engine/training.py”, line 1104, in fit
callback_metrics=callback_metrics)
File “/afs/in2p3.fr/home/s/sbilokin/.local/lib/python2.7/site-packages/keras/engine/training.py”, line 842, in _fit_loop
callbacks.on_epoch_end(epoch, epoch_logs)
File “/afs/in2p3.fr/home/s/sbilokin/.local/lib/python2.7/site-packages/keras/callbacks.py”, line 40, in on_epoch_end
callback.on_epoch_end(epoch, logs)
File “/afs/in2p3.fr/home/s/sbilokin/.local/lib/python2.7/site-packages/keras/callbacks.py”, line 296, in on_epoch_end
self.model.save(filepath, overwrite=True)
File “/afs/in2p3.fr/home/s/sbilokin/.local/lib/python2.7/site-packages/keras/engine/topology.py”, line 2427, in save
save_model(self, filepath, overwrite)
File “/afs/in2p3.fr/home/s/sbilokin/.local/lib/python2.7/site-packages/keras/models.py”, line 56, in save_model
model.save_weights_to_hdf5_group(model_weights_group)
File “/afs/in2p3.fr/home/s/sbilokin/.local/lib/python2.7/site-packages/keras/engine/topology.py”, line 2476, in save_weights_to_hdf5_group
dtype=val.dtype)
File “/afs/in2p3.fr/home/s/sbilokin/.local/lib/python2.7/site-packages/h5py/_hl/group.py”, line 108, in create_dataset
self[name] = dset
File “_objects.pyx”, line 54, in h5py._objects.with_phil.wrapper (/scratch/pip_build_sbilokin/h5py/h5py/_objects.c:2513)
File “_objects.pyx”, line 55, in h5py._objects.with_phil.wrapper (/scratch/pip_build_sbilokin/h5py/h5py/_objects.c:2466)
File “/afs/in2p3.fr/home/s/sbilokin/.local/lib/python2.7/site-packages/h5py/_hl/group.py”, line 277, in __setitem__
File “_objects.pyx”, line 54, in h5py._objects.with_phil.wrapper (/scratch/pip_build_sbilokin/h5py/h5py/_objects.c:2513)
File “_objects.pyx”, line 55, in h5py._objects.with_phil.wrapper (/scratch/pip_build_sbilokin/h5py/h5py/_objects.c:2466)
File “h5o.pyx”, line 202, in h5py.h5o.link (/scratch/pip_build_sbilokin/h5py/h5py/h5o.c:3726)

—————————————————–
The file names are unique, probably there is a collision of names in keras or h5py for me.

• Sviat September 11, 2016 at 10:15 am #

Found my mistake, I didn’t edit topology.py in keras correctly to fix another problem.

This is my second attempt to study NN, but I always have problems with versions, errors, dependencies and this scares me away.
For example now I have a problem to load the weights, using the example above on python 3 with intermediate weight files:

Traceback (most recent call last):
File “bot.py”, line 49, in
File “.local/lib/python3.3/site-packages/keras/engine/topology.py”, line 2490, in load_weights
File “.local/lib/python3.3/site-packages/keras/engine/topology.py”, line 2533, in load_weights_from_hdf5_group
weight_names = [n.decode(‘utf8’) for n in g.attrs[‘weight_names’]]
File “h5py/_objects.pyx”, line 54, in h5py._objects.with_phil.wrapper (/scratch/pip_build_/h5py/h5py/_objects.c:2691)
File “h5py/_objects.pyx”, line 55, in h5py._objects.with_phil.wrapper (/scratch/pip_build_/h5py/h5py/_objects.c:2649)
File “/.local/lib/python3.3/site-packages/h5py/_hl/attrs.py”, line 58, in __getitem__
attr = h5a.open(self._id, self._e(name))
File “h5py/_objects.pyx”, line 54, in h5py._objects.with_phil.wrapper (/scratch/pip_build_/h5py/h5py/_objects.c:2691)
File “h5py/_objects.pyx”, line 55, in h5py._objects.with_phil.wrapper (/scratch/pip_build_/h5py/h5py/_objects.c:2649)
File “h5py/h5a.pyx”, line 77, in h5py.h5a.open (/scratch/pip_build_/h5py/h5py/h5a.c:2179)
KeyError: “Can’t open attribute (Can’t locate attribute)”

• Jason Brownlee September 12, 2016 at 8:26 am #

I don’t think I can give you good advice if you are modifying the Keras framework files.

Good luck!

4. Alex September 14, 2016 at 11:46 pm #

Hi Jason,

I followed your tutorial and then built my own sequence-to-sequence model, trained at word level.
Might soon share results and code, but wanted first to thank you for the great post, helped me a lot getting started with Keras.

Keep up the amazing work!

5. Nader September 22, 2016 at 2:15 am #

When I try to get the unique set of characters:
I get the following:
[‘\n’, ‘ ‘, ‘!’, ‘”‘, “‘”, ‘(‘, ‘)’, ‘*’, ‘,’, ‘-‘, ‘.’, ‘:’, ‘;’, ‘?’, ‘[‘, ‘]’, ‘_’, ‘a’, ‘b’, ‘c’, ‘d’, ‘e’, ‘f’, ‘g’, ‘h’, ‘i’, ‘j’, ‘k’, ‘l’, ‘m’, ‘n’, ‘o’, ‘p’, ‘q’, ‘r’, ‘s’, ‘t’, ‘u’, ‘v’, ‘w’, ‘x’, ‘y’, ‘z’, ‘\xbb’, ‘\xbf’, ‘\xef’]

Note, that the ‘\r’ is missing, why ?

Thank you

• Jason Brownlee September 22, 2016 at 8:20 am #

It is not needed on some platforms, like Unix and friends. Only windows uses CRLF.

• Nader September 22, 2016 at 11:44 am #

I am using windows.
I am running the fitting as I type.
🙂

• Jason Brownlee September 22, 2016 at 5:29 pm #

Great!

• Nader September 22, 2016 at 11:28 pm #

Question:

Can I generate a text book the size of Alice in Wonderland using the same technique ?

And if so, how ?

Do I generate 50,000 characters for example ?

And how do I use a “SEED” to actually generate such a text ?

In the example you are using a 100 characters as a way to breakup the text and expose it to the network. DOES increasing the characters help with producing a more meaningful text ?

And another question ?

How many epochs should I run the fitting ? 100, 200 ? Because the loss keeps decreasing, but if it gets close to zero is that a good thing ?

Sorry for so many questions.

Thank you VERY VERY MUCH 🙂

• Jason Brownlee September 23, 2016 at 8:27 am #

Great questions, but these are research questions.

You would have to experiment and see.

• Gustavo Führ March 21, 2017 at 5:49 am #

I did a similar project, and I removed non ASCII characters using:

ascii_values = ascii_values[np.where(ascii_values < 128)]

6. Landon September 23, 2016 at 2:57 am #

Hey, nice article. Can you explain why you are using 256 as your output dimension for LSTM? Does the reasoning for 256 come from other parameters?

• Jason Brownlee September 23, 2016 at 8:29 am #

The network only outputs a single character.

There are 256 nodes in the LSTM layers, I chose a large number to increase the representational capacity of the network. I chose it arbitrarily. You could experiment with different values.

7. Lino October 24, 2016 at 3:15 pm #

awesome explanation. Thanks for sharing

• Jason Brownlee October 25, 2016 at 8:21 am #

I’m glad you found it useful Lino.

8. Shamit October 27, 2016 at 7:18 am #

Thanks for the nice article. Why do we have to specify the number of times-teps beforehand in LSTM ( in input_shape)? What if different data points have different number of sequence lengths? How can we make keras lstm work in that case?

• Jason Brownlee October 27, 2016 at 7:49 am #

Hi Shamit,

The network needs to know the size of data so that it can prepare efficient data structures (backend computation) for the model.

If you have variable length sequences, you will need to pad them with zeros. Keras offers a padding function for this purpose:
https://keras.io/preprocessing/sequence/

9. Julian October 27, 2016 at 6:50 pm #

Hi Jason

Thanks a lot for the code and the easy explanations.

Only a short notice: For the model checkpoints you will need the h5py module which was not preinstalled with my python. I obviously only noticed it after the first epoch, when keras tried to import it. Might be a good idea for people to check before they waste time running the code on slow hardware like I did 🙂

10. Bruce November 13, 2016 at 7:07 pm #

Hi Jason

I ran the exact same code of your small LSTM network on the same book(wonderland.txt) for 20 epochs. But the generated text in my case is not making any sense. A snippet of it is as below:

Seed:
” e chin.

æiÆve a right to think,Æ said alice sharply, for she was beginning to
feel a little worried ”

up!uif!tbccju!xp!cf!b!mpsu!pg!uif!tbbufs!bos!uif!ebsufs-!boe!uifo!tif
xbt!bom!uif!ebuufs!xiui!uif!sbtu!po!uif!!boe!uif!xbt!bpmjoh!up!uif!
bpvme-

uif!nbuuf!xbt!b!mpsumf!ujfff!up!cf!b!mpsu!pg!uif!tbbufs!bos!uif!ibouf
uif!tbt!pg!uif!xbt!tp!tff!pg!uif!xbt!pp!bo!bo!pg!tifof!!bou!uif!xbt!bom
!tif!ibe!tv!co!bo!uif!sbccju!xpvme!cf!bspmpvt!up!tffff!
uif!ibe!tp!cfe

Could you give some insights on why this is happening and how to fix it?

• Jason Brownlee November 14, 2016 at 7:43 am #

Ouch, something is going on there Bruce.

Perhaps confirm Keras 1.1.0 and TensorFlow 0.10.

It looks like maybe the network was not trained for long enough or the character conversion from ints is not working out.

Also, conform that there were no copy-paste errors.

• Rahul Kulhalli April 10, 2018 at 11:34 pm #

Hey Bruce, any updates on this? I tried using one-hot vectors as inputs instead of the ones mentioned in the post. I’m getting a similar output to yours. Something along the lines of this:

xxw’?,p?9l5),d-?l?sxwx?fbb?flw?g5ps-up ?’xx?,)lqc?lrex?fqp,)xw?gfu-fwf ,,x?up ?bxvcxexw?)fwgxg?,pq54d-;wp9?ql5),[?p ,?p;?,)x?;pwwfs)-lq?bp’x?l’

11. Keerthimanu November 24, 2016 at 10:19 am #

How to take the seed from user instead of program generating the random text ?

• Jason Brownlee November 24, 2016 at 10:45 am #

You can read from sys.stdin, for example:

• Srce Cde (Chirag) August 30, 2017 at 8:15 pm #

For user input you can use userinput = stdin.readline()

For ex the input is:

userinput = “Alice was beginning to get very tired of sitting by her sister on the bank, and of having nothing to”

sort_sen = sorted(list(p))
pattern = [char_to_int[value.lower()] for value in sort_sen]

This is how you can deal with user input. Hope this helps.

12. Victor November 27, 2016 at 6:28 pm #

Hi Jason,
Thank you so much for the great post!
Can you please explain the need to reshape the X ?
What is wrong with the initial shape of list of lists?
Thank you!

• Jason Brownlee November 28, 2016 at 8:42 am #

Hi Victor, great question.

All LSTM input must be in the form [samples, timesteps, features]. The data was loaded as [samples, timesteps]. Rather than timesteps, think of sequence – it’s the same thing.

I hope that helps.

13. ATM November 30, 2016 at 1:02 pm #

Using a similar approach: Can one generate numeric sequences from time series data, much like sentences? I don’t see a reason why we can’t. Any input is appreciated, Jason. thanks.

• Jason Brownlee December 1, 2016 at 7:15 am #

For sure, change the output from one node to n-nodes.

For sequence output, you’ll need a lot more data and a lot more training.

• ATM December 4, 2016 at 2:25 pm #

Thank you, but why should we change to n-nodes? Considering you generated a sequence of text, can’t I get a numeric sequence with the same 1 node setup?

Also, I don’t understand “index = numpy.argmax(prediction)
result = int_to_char[index]” part of the code. Can you please explain why its necessary?

I’m new to your website….Keep up the great work! Looking to hear from you.

• Jason Brownlee December 5, 2016 at 6:49 am #

Hi ATM,

Yes, there are a few ways to frame a sequence prediction problem. You can use a one-step prediction model many times. Another approach is to predict part or the entire sequence at once, and all the levels in between.

The line:

Selects the node with the largest output value. The index of the node is the same as the index of the class to predict.

I hope that helps.

• ATM December 7, 2016 at 5:31 pm #

Thanks for the clarification, Jason: I got the code running with decent predictions for time series data.

I think you might want to add in the tutorial that prediction usually works well for only somewhat statistically stationary datasets, regardless of training size?

I’ve tried it on both stationary and non-stationary, and I’ve come to this conclusion (which makes sense). Very often, time series collected from both financial and scientific datasets are not stationary, so LSTM has to be used very conservatively.

• Jason Brownlee December 8, 2016 at 8:15 am #

Thanks ATM! I agree, I’ll be going into a lot more details on stationary time series and making non-stationary data stationary in coming blog posts.

14. Sban December 6, 2016 at 2:16 am #

Hi jason,
I have been running this. But thebloss instead of decreasing is always increasing. So far, i ran it for 20 epochs. Do you have any idea what might be the case. I didn’t change anything in the program, though.

Best,
Sban

• Jason Brownlee December 6, 2016 at 9:53 am #

Maybe your network is too small? Maybe overlearning?

15. ben December 6, 2016 at 2:59 am #

hello,
how would you handle a much bigger dataset ? I scraped all public declarations of the French goverment for the past few years so my corpus is Total Characters: 465163150
Total Vocab: 146
the dataX and dataY will likely crash due to size constraints, how could I circumvent that issue and use the full dataset ?

• Jason Brownlee December 6, 2016 at 9:54 am #

Perhaps read the data in batches from disk – I believe Keras has this capability for images (data generator), perhaps for text too? I’m not sure off hand.

16. Julian January 2, 2017 at 5:10 pm #

Hi Jason,

Thanks for the example. I wonder about the loss function: categorical cross-entropy. I tried to find the source code but was not successful with it. Do you know what a loss of 1.2 actually means? Is there a unit to this number?

From my understanding, the goal is to somehow get the network to learn a probability distribution for the predictions as similar as possible to the one of the training data. But how similar is 1.2?

Obviously a loss of 0 would mean that the network could accurately predict the target to any given sample with 100% accuracy which is already quite difficult to imagine as the output to this network is not binary but rather a softmax over an array with values for all characters in the vocabulary.

Thanks again.

17. Eyal January 11, 2017 at 5:59 pm #

Hey, thanks for the post, will this work on Theano instead of tensorflow? I understand Keras can work with both, and I am having difficulties installing tensorflow on my mac

• Jason Brownlee January 12, 2017 at 9:25 am #

Yes, the code will work on either backend.

• Eyal January 14, 2017 at 1:01 am #

Thanks, TensorFlow worked after all. I am running the training on a big data like the book with these messages:
Total Characters: 237084
Total Vocab: 50
Total Patterns: 236984

on Epoche 13, loss is 1.5. I had to stop it there since it took my MacBook Pro 20 hours to create 13 epoches. Anyways, this is my result, notice it starts to repeat itself at a certain point:

is all the love and the sain on my baby

bnd i love you to be a line to be a line

to be a line to me, i’m she way the lov

es me down and i do and i do and i do an

d i do and i do and i do and i do and i

do and i do and i do and i do and i do a

nd i do and i do and i do and i do and i

do and i do and i do and i do and i do

and i do and i do and i do and i do and

This is another result from another run of the program:

oe mine i want you to ban tee the sain o

f move and i do and i do and i love you

to be a line to me, i’m she way the way

the way the way the way the way the way

the way the way the way the way the way

the way the way the way the way the way

the way the way the way the way the way

the way the way the way the way the way

This is pretty frustrating (while still incredibly awesome), do you have an idea why this should happen?

Regardless, this is such a great post that gives access to RNN and LSTM in a great nature!

18. Jack January 15, 2017 at 8:15 am #

Would we get better result if we trained the network at the word level instead of character? We would split the dataset into words, index them, the same way we do with the characters, and use them as input. Wouldn’t that at least eliminate the made up words in the generated text making it seem more plausible?

• Jason Brownlee January 16, 2017 at 10:35 am #

Yep. Try it out Jack! I’d love to see the result.

• Amin February 2, 2019 at 2:54 am #

I agree that the word level modeling would be much more intuitive and sentences might show some meaningful results.
One issue is that you have 47 features to choose from if you go character level. However, the number of unique words is orders of magnitude larger than unique characters. Therefore, your network becomes much larger. You can add an embedding to reduce input dimensionality, however, the output softmax will still be large.

19. Fateh January 18, 2017 at 12:51 am #

Is there any benchmark dataset for this task, to actually evaluate the model ?

• Jason Brownlee January 18, 2017 at 10:15 am #

Not as far as I’m aware Fateh.

20. Don January 20, 2017 at 10:34 am #

Hi Jason amazing post! I have a doubt. I tried the following: instead of training with Alice’s Adventures, I train with this a list of barcodes. Unique barcodes in a plain text (example, 123455, 143223, etc etc) and they can be of different lengths. Is this approach still valid? What i want to do is to input a potentially barcode with errors (maybe a character is missing) and the LSTM returns me “the correct one”. Any suggestion? Many thanks in advance!

• Jason Brownlee January 21, 2017 at 10:22 am #

Nice idea Don.

LSTMs can indeed correct sequences, but they would need to memorize all of the barcodes. You would then have to train it to fill in the gaps (0) with the missing values.

Very nice idea!

You could also try uses an autoencoder to do the same trick.

I’d love to hear how you go.

• Don January 21, 2017 at 10:58 pm #

Thanks Jason! For sure i will come back 🙂

Do you have any suggested read for either LSTM as a “word corrector” or an autoencoder for that task?

Best!

• Jason Brownlee January 22, 2017 at 5:12 am #

Sorry I don’t Don. Good luck with your project!

21. Alex Bing January 20, 2017 at 8:56 pm #

Hi Jason, thank you very much for your post.

I was thinking about changing the input and output to be a coordinate, which is 2D vector (x and y position) to predict movement.

I understood that we can change the input to be a vector instead of scalar. For example using one hot vector like you’re suggesting.

However, I don’t understand what to do when we want the output to be a 2D vector. Does this mean that I don’t need to use softmax layer as the output?

• Jason Brownlee January 21, 2017 at 10:29 am #

Hi Alex, interesting idea – movement generator.

You can output a vector by changing the number of neurons in the output layer.

22. Ashok Kumar January 28, 2017 at 3:34 am #

Hi Jason,

The features in the exercise are characters which are mapped to integers. It’s like replacing a nominal vector by a continuous variable. Isn’t that an issue?

Also, would you consider using Embedding as the first layer in this model – why or why not.

• Jason Brownlee January 28, 2017 at 7:53 am #

Hi Ashok,

The input data is ordinal, either as chars or ints. The neural net expects to work with arrays of numbers instead of chars, so we use ints here.

An embedding layer would create a projection for the chars into a higher dimensional space. That would not be useful here as we are trying to learn and generalize the way sequences of chars are put together.

I could be wrong though, try it and see if you can make it work – perhaps with projections of word sequences?

• Ashok January 28, 2017 at 8:34 am #

On the embedding layer, I have another question – how are the projections learnt? Is it a simple method like multiplying with a known matrix or are they learnt iteratively? If iteratively, then are they learnt a) first before the other weights or b) are the embedding projection weights learnt along with the other weights.

Ashok

• Ashok January 29, 2017 at 3:45 am #

Jason:

The below code for text generation is treating chars as nominal variables and hence giving assigning a separate dimension for each. Since this is a more complex approach, I will check if this leads to an improvement.

Again, thanks for your response. It only added to my curiosity .

https://github.com/fchollet/keras/blob/master/examples/lstm_text_generation.py

• Jason Brownlee February 1, 2017 at 10:11 am #

I’d love to hear how the representational approaches compare in practice Ashok.

• Konstantin March 16, 2017 at 7:55 am #

Jason, could you explain please how the input data is ordinal? Instead of one-hot encoding we simply enumerate characters.

• Pasquinell August 2, 2017 at 9:18 pm #

I have the same question. Have you found any answer?

• Jason Brownlee August 3, 2017 at 6:51 am #

If your input is text, you can use:

– an integer encoding of the sequence
– a bag of word encoding (e.g. counts of tf-idf)
– one hot encoding of integer encoding
– word embedding (word2vec, glove or learned)

Does that help?

23. Hendrik March 3, 2017 at 8:24 pm #

Hello Jason!

Nice work. But I don’t really understand what is the point of applying RNN for this particular task. If one wants to predict the next character of a 100-long sequence, this is seems me achievable with any other common regression task where you feed in the training sequences and the following single character as output while training. What is the additional feat of your approach?

• Jason Brownlee March 6, 2017 at 10:47 am #

Yes, here are many ways to approach this problem.

LSTMs are supposed to particularly good at modeling the underlying PDF of char or words in a text corpus.

24. Hendrik March 7, 2017 at 9:06 pm #

The example doesn’t run anymore under TensorFlow 1.0.0. (previous versions are OK, at least with 0.10.0rc0).

25. Gustavo Führ March 21, 2017 at 6:01 am #

Great post Jason, I was trying to do almost the same thing and your post gave a lot of help. As some of the comments suggest, the RNN seems to achieve a loop quite quickly. I was wondering what you meant with “Add dropout to the visible input layer and consider tuning the dropout percentage.”, seems it appears to attack this problem.

• Jason Brownlee March 21, 2017 at 8:44 am #

You can apply regularization to the LSTMs on recurrent and input connections.

I have found adding a drop out of 40% or so on input connections for LSTMs very useful on other projects.

More details here:
https://keras.io/layers/recurrent/#lstm

• Gustavo Führ March 23, 2017 at 9:31 am #

Actually Jason,

I advanced my experiments and found some interesting things, I will probably do a Medium about it.

For the loop problem, as mentioned in http://karpathy.github.io/2015/05/21/rnn-effectiveness/, makes no sense to always use the argmax in the generation. Since the network output a list of probabilities is easy to randomically sample letter using this distribution:

def sample_prediction(char_map, prediction):
rnd_idx = np.random.choice(len(prediction), p=prediction)
return char_map[rnd_idx]

That simple change made the network to avoid loops and be much more diverse. (Using
temperature as you mentioned will be even better)

Here is a sample of a one-layer network (same as yours), trained with Cervante’s Dom
Quixote, for 20 epochs:

=====
from its beginning to its end, and above all manner; as so able to be almost unable to repeat the, which he had protected him to leave it. As now, for her face because I will be, that he ow her its missing their resckes my eyes, proved
for the descance in the mouth as saying, to eat for a
shepherdess and this knight ais, they
did so that was rockly than to possess the new lose of, that in a sword of now another
golden without a change on which not this; if shore of the Micomicona his dame, ‘To know some little why as capacquery; I will make any ammorimence is seen in their
countless
soicour and the adventure of her damsel a pearls that shows and vonquished callshind him, away itself, her fever and evil comrisoness by where they
will show not in time in which all the chain
himself of the
solmings of the hores of this
should have punisented, not to portion for, as it
just be-with his sort rich as the shaken of the
sun to

Am I he will not believe. I will disreel her more so knight, for
=====

Isn’t that cool? I hope that helps other people.

• Jason Brownlee March 24, 2017 at 7:51 am #

Very cool Gustavo, thanks for sharing!

• Jatin March 31, 2017 at 10:12 am #

First of all big thanks to Jason for such a valuable write up.

I have 1 question :

To Gustavo Führ and Jason Brownlee

Could you please explain in a little more detail what you did there. i didn’t understand, how you got such a good output on a single layer.

I ran my training on this book as my data -> http://www.gutenberg.org/cache/epub/5200/pg5200.txt

See below the output i am getting..

Seed :
see it coming. we can’t all work as hard as we have to and then come hometo be tortured like this, we can’t endure it. i can’t endure it anymore.” and she broke out so heavily in tears that they flo
Prediction :
led to het hemd and she was so the door and she was so the door and she was so the door and she was so the door and she was so the door and she was so the door and she was so the door and she was so the door and she was so the door and she was so the door and she was so the door and she was so the door and she was so the door and

As you can see here a text is repeating itself. This the result of 20th epoch out of 50 epochs in a dual layer LSTM with 256 cells and a sequence of 200 characters.

This is a result pattern i am seeing in all my trainings and i believe i am doing something wrong here. Could you please help out?

• Luke September 30, 2019 at 3:03 pm #

According to code above it repeats because of the MOST probable (argmax) on the line:
index = numpy.argmax(prediction)
and it should rather be:
index = numpy.random.choice(len(prediction[0]),p=prediction[0])

Now this works but it is giving gobeldegook instead of semi-english repeating words. Perhaps I need to train for hours longer and have loss <2.5?

• Rob November 23, 2017 at 6:24 am #

I had a very similar experience in my own experimentation. This comment verified what I had found — thank you for sharing Jatin!

26. Henok S Mengistu April 5, 2017 at 8:32 am #

what is the need for normalization

# normalize
X = X / float(n_vocab)

• Jason Brownlee April 9, 2017 at 2:32 pm #

This normalizes the integer values by the largest integer value. All values are rescaled to 0-1.

• Eric September 12, 2017 at 4:59 pm #

Is this because it’s using a relu activation function? Or because generally you need input values to be between 0-1?

• Jason Brownlee September 13, 2017 at 12:29 pm #

LSTMs prefer normalized inputs – confirm this with some experiments (I have).

27. June April 11, 2017 at 11:20 pm #

Hello~ I am your one of fans.

I have a question about lstm settings that I may apply your text generation model in a different way.

Do you think it is possible to generate texts by adding a word rather than a character.

If it is possible, adding word_embedding layer is effective for a performance of text generation??

• Jason Brownlee April 12, 2017 at 7:53 am #

Yes, I expect it may even work better June. Let me know how you go.

• JUNETAE KIM April 12, 2017 at 12:15 pm #

Thanks, Jason.

I will try it~!

You know, I am a doctoral student majored in management information system.

My research interest include medical informatics and healthcare business analytics.

Your excellent blog posts have helped me a lot.

I appreciated it so much!

• Jason Brownlee April 13, 2017 at 9:54 am #

I’m really glad to hear that. Thanks for your support and your kind words, I really appreciate it.

28. Dave May 6, 2017 at 11:43 am #

Hi,

Very interesting example – can’t wait to try training on other texts!

I have a question regarding the training of this, or in fact any neural network, on a GPU – I have a couple of CNNs written but that I can’t execute :(.

Suppose I am using either Amazon EC2 or Google Cloud, I can successfully log into the instance and run a simple ANN using just a CPU but I am totally confused as to how to get the GPU working. I am accessing the instance from Windows 10. Can you tell me the exact steps I need to do – presumably I need to get CUDA and CUDNN somehow? Then is there any other stuff I need to do or can I just pip install the necessary packages and then execute my code?

Thank you so much.

29. FractalProb May 15, 2017 at 2:41 pm #

nice post!
two questions,
1) what is the role of “seq_in = [int_to_char[value] for value in pattern]” in line 63 (doesn’t seem to be used in the loop) and,
2) could you expand on where and how Gustavo’s fix is needed? (I also get repeated predictions using argmax)

Does your Deep Learning book expand on (2) in regards to LSTM’s?
It seems problematic to output the same predictions over different inputs.

• Jason Brownlee May 16, 2017 at 8:37 am #

seq_in is the input sequence used in prediction. Agreed it is not needed in the latter part of the example.

His fix is not needed, it can just reduce the size of the universe of possible values. My book does not go into more detail on this.

I am working on a new book dedicated to LSTMs that may be of interest when it is released (next month).

30. FractalProb May 16, 2017 at 3:54 am #

Here is my implementation of Gustavo’s suggestions so that the predicted output tends to avoid repeated patterns

def sample_prediction(prediction):
“””Get rand index from preds based on its prob distribution.

Params
——
prediction (array (array)): array of length 1 containing array of probs that sums to 1

Returns
——-
rnd_idx (int): random index from prediction[0]

Notes
—–
Helps to solve problem of repeated outputs.

len(prediction) = 1
len(prediction[0]) >> 1
“””
X = prediction[0] # sum(X) is approx 1
rnd_idx = np.random.choice(len(X), p=X)
return rnd_idx

for i in range(num_outputs):
x = np.reshape(pattern, (1, len(pattern), 1))
x = x / float(n_vocab)
prediction = model.predict(x, verbose=0)
#index = numpy.argmax(prediction)
# per Gustavo’s suggestion, we should not use argmax here
index = sample_prediction(prediction)
result = int_to_char[index]
#seq_in = [int_to_char[value] for value in pattern]
# not sure why seq_in was here
sys.stdout.write(result)
pattern.append(index)
pattern = pattern[1:len(pattern)]
print “\nDone.”

31. Rathish May 16, 2017 at 5:01 am #

Hi Jason, great post.

I wanted to ask you if there is a way to train this system or use the current setup with dynamic input length.

The reason for that is if one to use this for real text generation, given a random seed of variable length say.. “I have a dream” or “To be or not to be” would it be possible to still generate coherent sentences if we train using dynamic length?

I tried with pad sequence in the predict stage (not the train stage) to match the input length for shorter sentences, but that doesn’t seem to work.

Regards

• Jason Brownlee May 16, 2017 at 8:51 am #

Yes, you are on the right track.

I would recommend zero-padding all sequences to be the same length and see how the model fairs.

You could try Masking the input layer and see what impact that has.

You could also try truncating sequences.

I have a post scheduled that gives many ways to handle input sequences of different lengths, perhaps in a few weeks.

32. Alex May 18, 2017 at 6:48 pm #

Hi,

seq_in = raw_text[i:i + seq_length]
seq_out = raw_text[i + seq_length]

should model predict next one abc -> d instead of abc -> c

33. abbey May 23, 2017 at 1:40 am #

Hi,

I need to thank you for all the good work on Keras, wounderful and awesome package.

However, i got a floating point exception (core dumped) running the code. Please, i need your advise to resolve the issues. I have upgrade the my keras from 1.0.8 to 2.0.1 and the issues is still the same.

88/200 [============>……………..] – ETA: 27s – loss: 11.3334 – acc: 0.0971{‘acc’: array(0.10470587760210037, dtype=float32), ‘loss’: array(9.283703804016113, dtype=float32), ‘batch’: 88, ‘size’: 17}
89/200 [============>……………..] – ETA: 26s – loss: 11.3103 – acc: 0.0972Floating point exception (core dumped)

Warm Regards

• Jason Brownlee May 23, 2017 at 7:54 am #

I’m sorry to hear that.

Perhaps you could try a different backend (theano or tensorflow)?

Perhaps you could try posting to stackoverflow or the keras user group?

• abbey May 23, 2017 at 6:23 pm #

Thank you Jason.

34. abbey May 23, 2017 at 9:20 pm #

Hi guys,

With Tensorflow backend, i got different error messages. After about 89 batch from the second epochs, the loss become nan and the accuracy also the same. Any suggestion or advices

87/200 [============>……………..] – ETA: 113s – loss: 9.4303 – acc: 0.0947{‘acc’: 0.10666667, ‘loss’: 9.2033749, ‘batch’: 87, ‘size’: 32}
88/200 [============>……………..] – ETA: 112s – loss: 9.4277 – acc: 0.0949{‘acc’: 0.10862745, ‘loss’: 9.2667055, ‘batch’: 88, ‘size’: 17}
89/200 [============>……………..] – ETA: 110s – loss: 9.4259 – acc: 0.0950{‘acc’: nan, ‘loss’: nan, ‘batch’: 89, ‘size’: 0}
90/200 [============>……………..] – ETA: 108s – loss: nan – acc: nan {‘acc’: nan, ‘loss’: nan, ‘batch’: 90, ‘size’: 0}
91/200 [============>……………..] – ETA: 106s – loss: nan – acc: nan{‘acc’: nan, ‘loss’: nan, ‘batch’: 91, ‘size’: 0}
92/200 [============>……………..] – ETA: 105s – loss: nan – acc: nan{‘acc’: nan, ‘loss’: nan, ‘batch’: 92, ‘size’: 0}

35. Piush May 24, 2017 at 11:31 pm #

Thanks for the example. Piush.

36. abbey May 26, 2017 at 3:55 pm #

Hi Jason,

I tried your suggestion and the problem of nan for loss and accuracy still the same after second epoch while the batch size is 39/50.

I have also try all the activation function and regularization but still the same problems. Too bad!

37. abbey May 30, 2017 at 4:43 pm #

Hi Jason,

The problem of loss and accuracy becoming nan after few epoch as to do with the batch generator. I fix it now.

Thank you so much.

38. IanEden June 1, 2017 at 8:53 am #

Hello Jason,

I was wondering how I could add more hidden layers. I was wondering if this could help generate text with more efficiency

I tried using model.add(LSTM(some_number)) again but it failed and gave me the error:

“Input 0 is incompatible with layer lstm_3: expected ndim=3, found ndim=2”

39. Mo June 8, 2017 at 8:59 am #

Hello Jason,

Thanks for your great post it helped me a lot. Since I finished reading your post, I was thinking of how to implement it in a word level instead of character level. I am just confused of how to implement it because with characters we only have few characters but with words we might have say 10000 or even more. Would you please share your thoughts. I am really excited to see the results in a words-level and make further enhancements.

• Jason Brownlee June 9, 2017 at 6:16 am #

Great idea.

I would recommend ranking all words by frequency the assign integers to each word based on rank. New words not in the corpus can then also be assigned integers later.

40. Adi June 16, 2017 at 12:27 am #

hi jason , thanks for such an awesome post
I keep getting the error :
ImportError: load_weights requires h5py.

• Jason Brownlee June 16, 2017 at 8:04 am #

Perhaps your environment cannot see the h5py library?

Confirm you installed it the correct way for your environment.

Try important the library itself and refine env until it works.

Let me know how you go.

• Adi June 18, 2017 at 5:07 pm #

It was solved, I just restarted python after installing h5py.
when I predicted using the model, it through the characters like this, whats up with the 1s why are they coming?
t1
h1
e1
1
s1
e1
r1
e1
1
t1
o1
1
t1
e1
e1
1
w1
e1
r1
1
i1
n1
1
a1
n1
d1

• Jason Brownlee June 19, 2017 at 8:34 am #

I don’t know. Confirm that you have a copy of the code from the tutorial without modification.

• Adi June 20, 2017 at 2:10 pm #

can we use print() in place of sys.stdout.write()

• Jason Brownlee June 21, 2017 at 8:08 am #

Sure.

41. Abbey June 21, 2017 at 9:09 pm #

Hi Jason,

Please, I need your help. I try to validate my test data using the already training result from the model but I got the following error:

##################
AttributeError Traceback (most recent call last)
in ()
1 # load weights into new model
3 predictions = model.predict([test_q1, test_q2, test_q1, test_q2,test_q1, test_q2], verbose = True)

AttributeError: ‘History’ object has no attribute ‘load_weights’

In [ ]:
#####################

Below is the snippets of my code:

###Fitting model

save the best weights for predicting the test question pairs

save_best_weights = “weights-pairs1.h5”

checkpoint = ModelCheckpoint(filepath, monitor=’loss’, verbose=1, save_best_only=True, mode=’min’)

callbacks = [ModelCheckpoint(save_best_weights, monitor=’val_loss’, save_best_only=True),
EarlyStopping(monitor=’val_loss’, patience=5, verbose=1, mode=’auto’)]
start = time.time()
model_info=merged_model.fit([x1, x2, x1, x2, x1, x2], y=y, batch_size=64, epochs=3, verbose=True,
validation_split=0.33, shuffle=True, callbacks=callbacks)
end = time.time()
print(“Minutes elapsed: %f” % ((start – end) / 60.))

#####evaluting

#load weights into a new model

predictions = model.predict([test_q1, test_q2, test_q1, test_q2,test_q1, test_q2], verbose = True)

Rgeards

42. Kunal chakraborty July 2, 2017 at 8:56 pm #

Hello Jason,

Great tutorial. I have just one doubt though, in np.reshape command what does feature mean?
and why is it set to 1?

• Jason Brownlee July 3, 2017 at 5:32 am #

The raw data are sequences of integers. There is only one observation (feature) per time step and it is an integer. That is why the first reshape specifies one feature.

43. Alex July 4, 2017 at 12:35 am #

Excellent tutorial!

I have one question. I need predict some words inside text and I currently use LSTM based on your code and binary coding. Is a good practice to use n-words behind and k-words ahead of word which I want to predict?
It’s good to train model on data like this or we must rather use only data behind our prediction?

• Jason Brownlee July 6, 2017 at 10:03 am #

Try a few method and see what works best on your problem.

44. Francesca July 11, 2017 at 3:49 pm #

Thanks, This can make a good reference for me

45. Joe Melle August 5, 2017 at 6:20 am #

Why not one hot encode the numbers?

• Jason Brownlee August 6, 2017 at 7:28 am #

Sure, try it.

Also try an encoding layer.

There are many ways to improve the example Joe, let me know how you go.

46. Sivasailam September 3, 2017 at 8:32 pm #

Jason,
Great post. Thanks much. I am very new to LSTM. I am trying to recreate the codes here. How to increase the number of epochs?
Thanks

• Jason Brownlee September 4, 2017 at 4:31 am #

Change the value after “epochs=”

47. Greg September 5, 2017 at 4:58 pm #

Hi Jason,

Thank you so much for these guides and tutorials! I’m finding them to be very helpful.

• Jason Brownlee September 7, 2017 at 12:42 pm #

I’m glad to hear that Greg.

48. Don September 7, 2017 at 11:04 am #

Hi Jason,

Thanks for your great post, and all the other great post as a matter of fact. I’m interested to train the model on padded sentences rather than random sequences of characters, but it’s not clear to me how to implement it. Can you please elaborate about it and give an example?

Many thanks!
Don

49. Don September 8, 2017 at 3:10 pm #

Thanks again for the post and all the info! I have another question. What should I do if I want to predict by a inserting a string shorter than 100 characters?

Many thanks again!

50. Don September 10, 2017 at 2:55 am #

Thanks!

I have a question regarding the number of parameter in the model and the amount of data. When I look at the summary of the simplest model, I get: Total params: 275,757.0, Trainable params: 275,757.0, and Non-trainable params: 0.0 (for some reason I didn’t succeed to sent a reply with the whole summary).

The number of characters in the Alice book is about 150,000. Thus, isn’t the number of parameter larger than the number of characters (data)?

Thanks again!

• Jason Brownlee September 11, 2017 at 12:02 pm #

Yes.

• Don September 11, 2017 at 2:12 pm #

Isn’t that over-fitting? I ask because you suggested to improve the quality of results by developing an even much larger LSTM network. But if in the simple LSTM network you already have more parameters than data, shouldn’t you simplify the network even more?

Thanks a lot for all the tips!

• Jason Brownlee September 13, 2017 at 12:21 pm #

A simpler model is preferred, but overfitting is only the case when skill on test/validation data is worse than train data.

51. Manar September 13, 2017 at 4:32 am #

Thanks Dr.Jason,

What if I want to output the probability of a sequence under a trained model rather than finding the most probable next charterer.

The use case that I have in mind is to feed the model test sentence and it prints the probability of this sentence being true under the model

• Jason Brownlee September 13, 2017 at 12:34 pm #

That would be a different framing of the problem, or perhaps you can collect the probabilities of each char step by step.

• Manar September 14, 2017 at 2:09 am #

So when I train, Xs is similarly he sequence, but what would the Ys be ?
Thanks!

• Jason Brownlee September 15, 2017 at 12:07 pm #

The next char or word in the sequence.

• Manar September 15, 2017 at 12:38 pm #

If we do this then how it will differ from the original framing of the problem? I mean the resulted probability will measure the how likely it is a character X will be the next rather than the probability of the seen sequence under the trained model

e.g. if we feed the model with “I am” and the next word either [egg or Sam]. Following the above-reply the output will be the probability of egg or Sam to be the next. Rather, I need to find the probability of the entire segment ” I am Sam” and “I am egg” to be able to tell which one makes more sense

• Jason Brownlee September 16, 2017 at 8:35 am #

Yes, it is the same framing. The difference is how you handle the probabilities. Sorry, I should have been clearer.

E.g. you can beam search the probabilities across each word/char, or output the probabilities for a specific output sequence, or list the top n most likely output sequences.

Does that make sense?

52. Rohit September 18, 2017 at 7:19 pm #

Hi Jason,
Thanks for the tutorial. When I am using one-hot encoded input, my accuracy and loss is not improving and becomes stagnant

function:
def one_hot_encode(sequences, next_chars, char_to_idx):
X = np.zeros((n_patterns, seq_length, N_CHARS),dtype=float)
y = np.zeros((n_patterns, N_CHARS),dtype=float)
for i, sequence in enumerate(sequences):
for t, char in enumerate(sequence):
X[i, t, char] = 1
#print(i,t,char)
y[i,[next_chars[i]]] = 1
return X, y

X, y = one_hot_encode(dataX, dataY, char_to_int)

Is it the layer size or a bug in one_hot_encoding?

53. Vinicius September 20, 2017 at 1:07 am #

Hi, thanks for the great tutorial!

If I begin with some random character and use the trained model to predict the next one, how can the network generate different sentences using the same first character?

• Jason Brownlee September 20, 2017 at 5:59 am #

It will predict probabilities across all output characters and you can use a beam search through those probabilities to get multiple different output sequences.

54. Sam September 20, 2017 at 10:15 am #

Hello,
I’m attempting to run the lastf ull code example for generating text using the loaded LSTM model.
However, line 59 simply produces the same number (effectively a space character) each time across the whole 1000 range.
Not sure what I’m doing wrong ?

• Jason Brownlee September 20, 2017 at 3:01 pm #

Sorry to hear that Sam.

Have you tried to run the example a few times?
Have you confirmed that your environment is up to date?

55. Iain Strachan September 29, 2017 at 1:23 am #

I did the experiment with a different corpus of text; Shakespeare’s sonnets, two LSTM layers of 256, and 20 epochs (probably needs longer). I found that the method in your post where you simply choose the next character from the maximum output, quickly produces very repetitive text – the same two lines repeat after a while, like this:

Seed:
” raise is crowned,
but those same tongues that give thee so thine own,
in other accents do this p ”
oor so me
the world with the shee the world with thee,
the world with the shee the world wour self stolne,
the world with the shee the world with thee shee steel.

12
the world with the the world with the dearty see,
the world with the shee the world with thee wour self,
the world with the shee the world with thee shee shee,
the world with the shee the world wour self so bear,
the world with the shee the world with thee shee shee,
the world with the shee the world wour self so bear,
the world with the shee the world with thee shee shee,
the world with the shee the world wour self so bear,
the world with the shee the world with thee shee shee,
the world with the shee the world wour self so bear,
the world with the shee the world with thee shee shee,
the world with the shee the world wour self so bear,
the world with the shee the world with thee shee shee,
the world with the shee the world wour self so bear,
the world with t

However, if you introduce some randomness by sampling the prediction probability distribution randomly, you get much more interesting results – although it’s gibberish, there are many gibberish words that are pronounceable – ie not just randomly selected, and the overall effect looks like it might be middle English, or even german in places. The randomness means that sometimes it doesn’t get the line breaks in approximately the right place. Nonetheless a very interesting result, and it doesn’t loop.

I love the way it puts the sonnet number in the right place!

Seed:
” nds to the course of alt’ring things:
alas why fearing of time’s tyranny,
might i not then say ‘ ”
to sof bn bde,
the iroelge gatesds bever neagne brmions
ph puspper stais, delcs tien love ahena,
thich thi derter worndsm hin fafn’sianlu thee:
shese blunq so me for they fadr mnhit creet.

64
cscv thy swbiss io ht the yanjes sast,
mftt pwr thet thai, io mengdvr to mehpt
to slobn of wourt)s fron shy fort siou mavt dr chotert fold (mn heart oo shoedh stcete,
that hil bu toun fass rrunhng of aeteinen
rie trilg mf prinlenlss, and ge voids,
batse mi tey betine out le il your wou,
these own taref in formuiers wien,io hise)
h eaar nonlhng uhe wari her bfsriite,
ming oat,s al she wantero wo me.

96
theteeo i fave shee wout fnler fiselgct
sreanind,
whthsu doom that brss nn len; a
tekl of me
tr pfngslcs,tien gojeses tore dothen,
o beau aal wierefe, oo ttomnaei ofs.
au in doace fiasss eireen thae despered,
thv urut ninyak wtaprr’ and thereoo dlg hns
cooh,
afaonyt o

PS this is just to test it out – I really want to generate synthetic time-series data, but I found that just predicting the next value then running generatively always produced a decay to a constant value.

• Iain Strachan September 29, 2017 at 1:26 am #

PS – having cut pasted it the pasting on this blog puts the sonnet number in the wrong place at the beginning of the line. In the generated text, the sonnet number is centred, ie preceded by a number of spaces, as in the original.

• Jason Brownlee September 29, 2017 at 5:08 am #

Very nice, thanks for sharing.

Perhaps you can use a beam search to better sample the output probabilities and get a sequence that maximizes the likelihood.

• Iain Strachan September 29, 2017 at 8:07 pm #

Sounds like a good idea, and also applicable to my goal – to generate synthetic sensor data.

56. Vibhu October 15, 2017 at 7:41 am #

Hey Jason,

Thanks for such a nice effort in deep learning stuff.

I am trying to run your code on my machine and it throws me this hdf5 error:

f = h5py.File(filepath, mode=’r’)
File “C:\Users\CPL-Admin\AppData\Local\Programs\Python\Python36\lib\site-packages\h5py\_hl\files.py”, line 269, in __init__
fid = make_fid(name, mode, userblock_size, fapl, swmr=swmr)
File “C:\Users\CPL-Admin\AppData\Local\Programs\Python\Python36\lib\site-packages\h5py\_hl\files.py”, line 99, in make_fid
fid = h5f.open(name, flags, fapl=fapl)
File “h5py\_objects.pyx”, line 54, in h5py._objects.with_phil.wrapper
File “h5py\_objects.pyx”, line 55, in h5py._objects.with_phil.wrapper
File “h5py\h5f.pyx”, line 78, in h5py.h5f.open
OSError: Unable to open file (unable to open file: name = ‘weights-improvement-47-1.2219-bigger.hdf5’, errno = 2, error message = ‘No such file or directory’, flags = 0, o_flags = 0)

• Jason Brownlee October 16, 2017 at 5:39 am #

It looks like the file or path does not exist.

Perhaps check that the file with that name exists in the current working directory.

• Phil June 20, 2018 at 12:13 pm #

Thanks for the interesting post. I am getting the exact same errors as Vibhu. When you say working directory do you mean the one in which the code resides? In packages?

• Jason Brownlee June 21, 2018 at 6:04 am #

The directory where you are working, where you have saved your code file.

57. Flo October 15, 2017 at 8:02 am #

Hello.

I’m currently trying to convert your process, using characters as an unit of information, to a system where I’m using the words as an unit of information. Currently dabbling in the domain of NN, I’m trying to figure out why I would get a repeated sequence of words without any change, and I may have some clues:
– Too much classes? With letters we have arond 30 or so classes (as we have the alphabet + the residual tick/backtick). Does it have really an impact during learning so generation can give this kind of repetitive result?
+ A few notions from the Markov chain is staying with me and I have a supposition why the system repeats in such cases : maybe the link between words isn’t pronounced enough fror the generator to choose anything else than what was provided to it?
– Wrong parameters? I’m trying multiple settings for the learning process, such as changing seq_length to 10 or 50 or something else and/or changing the batch size?
– Corpus too small? Should I grab other books?
– Training too short? Is 60 epochs a relatively small or important processus?
– Note : I’m currently using a set to remove repeated sequences once the text tokenized . It really helps is seq_len is small, I think it could also help for character-based processing. That’s additional init overhead, but I think it might worth it.

I copied the code into this gist to avoid cluttering any further the comment log : https://gist.github.com/anonymous/28f30611bb0849ef0d99fd341e6e1d7b

Thanks for the article, it was pretty interesting and made me learn quite a bit on notions I forgot or didn’t understand well. I look forward to learn more and understand how to progress!

Have a nice day!

• Jason Brownlee October 16, 2017 at 5:40 am #

Great work, I recommend testing each hypothesis.

58. Bhau October 17, 2017 at 6:19 am #

Instead of letters, can we map each individual word to a number? Then we can use a few words worth of context to predict the next word.
So we could have [“I”,”am”,”a”] as the input and have [“human”] as the output.

I’m still learning a few concepts, so I might be wrong.

59. Stuart October 25, 2017 at 1:13 pm #

• Stuart October 25, 2017 at 1:16 pm #

nevermind, i’m totally wrong. please delete 🙂

• Jason Brownlee October 25, 2017 at 4:02 pm #

You may have to visit the link twice to set/use the cookie.

60. Ari November 14, 2017 at 1:21 pm #

Just wanted to say you are an amazing human who is enabling tons of people to learn complex material that otherwise might not be able to. I’m also astounded with your dedication in responding to literally every single comment, and within a short time too. You’re making the world a better place.

• Jason Brownlee November 15, 2017 at 9:46 am #

Thanks Ari, I appreciate your support and recognition!

61. Ying Sha November 20, 2017 at 12:11 am #

Hi,Jason, Thank you so much for the great post!
I ran the exact same code of your small LSTM network on the same book(wonderland.txt)
for 20 epochs.
But I get an error with this line:

The error:
“TypeError: Expected int32, got list containing Tensors of type ‘_Message’ instead.”

Could you please give me some insights on why this is happening and how to fix it?
Thank you very much!

Tensorflow: 0.10.0rc0
Keras: 2.1.1

• Jason Brownlee November 20, 2017 at 10:17 am #

Sorry, I have not seen this error before.

Perhaps double check that you have copied all of the code from the example exactly?

62. Martin November 30, 2017 at 10:25 pm #

Hello Jason,

I have been looking at examples of LSTM text generation for some time and I have the following question:

Once the network has been trained and all the weights have been fixed at some acceptable values, would the network generate different text every time it is fed the same input seed text (“What would happen” , for example), or would the output text sequence always be the same for that exact input seed because the next character with the highest probability after the sequence “What would happen” would always be predicted to be the same on each run due to the network’s fixed weights, and then the next one after appending the previously predicted character to the input as well, etc.?

I am working on a similar LSTM network in tensorflow for a sequence-labeling problem, and so far it appears that my generated output sequence is always exactly the same for a fixed starting input. Is there a way to generate different output sequences for the same input seed?

• Jason Brownlee December 1, 2017 at 7:34 am #

Once the network is trained, it will make deterministic predictions.

• Martin December 1, 2017 at 11:14 am #

Thank you for the answer. I suppose what I was describing was a stochastic Neural Network.

63. Martin December 7, 2017 at 3:40 am #

Hey Jason

May I ask why you used a normalized X vector as input and not just onehot-encoded input? Is this done by someone else as well? Just curious because I don’t really see an advantage over onehot encoding.

• Jason Brownlee December 7, 2017 at 8:07 am #

I would recommend one hot encoding instead these days.

64. Arshdeep December 13, 2017 at 6:10 pm #

Can you put your pretrained weights?

• Jason Brownlee December 14, 2017 at 5:34 am #

Yes, you can use pre-trained weights.

65. Emilia Stark December 13, 2017 at 6:11 pm #

Shouldn’t we use word2vec instead of one-hot encoding?

66. David December 31, 2017 at 8:38 am #

Hi Jason,
thanks for this great post. Do you have experience in using the LSTM as a model for Genetic Programming to further improve the output? Or have you ever heard about it? It would then be something like an EDA-GP. If you know anything about it, i’d be happy if you let me know!
David

• Jason Brownlee January 1, 2018 at 5:26 am #

I’ve not heard about using LSTMs with GPs. Interesting idea.

What makes you want to explore this combination?

• David January 3, 2018 at 4:00 am #

Oh I was just wondering. I’ve recently seen some other text generation work, they implemented it quite similarly as you did but came to the conclusion that in future this could probably be improved by using the LSTM as a model for a genetic program, to add further constraints and thus to improve the output. I just wanted to know if something like this has already been implemented. Yesterday I came across a really interesting paper. (https://arxiv.org/pdf/1602.07776.pdf) They invented a new method: Recurrent Neural Network Grammars (RNNGs). They report major improvements both in language modeling as also in parsing. I will maybe try this out, as a topic for my master thesis.
Best regards and happy new year!
David

• Jason Brownlee January 3, 2018 at 5:40 am #

Looks interesting, thanks for sharing.

67. Jens Albrecht January 8, 2018 at 11:00 pm #

Hi Jason, thank you for your great site and all the material that you offer.

I’ve been trying to modify the LSTM to predict words instead of letters. I stripped all special characters and numbers from the text and vectorized the words. It’s 26386 words in total and 2771 distinct words. The model could be trained, but it always and only predicts the word “the” as next word. I tried different sequence lengths with the simple and the complex model, but the result stays the same. Of “the” is the most frequent word with 1600 occurences, but shouldn’t the model still be able to predict something else? Or is the corpus just too small? What do you think?

68. kay January 11, 2018 at 6:27 pm #

Hi, Thanks for good contents, I have a question.
Is it possible to make similar sentence with RNN?
For example, seed sentence : “The rabbit-hole went straight on like a tunnel for some way”
generating several similar sentences :
” The rabbit is go straight on like a tunnel”
” The rabbit hole went straight”
“The rabbit straight for some way”

I don’t focus the semantic for sentences, just need to make training data with tagging to solve other text domain problem

• Jason Brownlee January 12, 2018 at 5:52 am #

Yes, generate multiple times, or generate once and use beam search to read off multiple output lines.

69. Prashant Goyal February 10, 2018 at 10:00 am #

This tutorial was a great help to me. Thanks for this. But I have a question regarding deep learning. As Deep Neural Networks take large amounts of time to train how can I tune different hyper-parameters of the model easily?

• Jason Brownlee February 11, 2018 at 7:50 am #

You can use multiple computers in the cloud in parallel to perform parameter tuning.

You could also try tuning using a smaller amount of data that requires less time to train. But this may impact the quality of the results.

• Prashant Goyal February 12, 2018 at 5:13 am #

Thanks, Jason.

70. Amit Moondra February 11, 2018 at 8:20 am #

Hi Jason, Thank you so much for this blog. It was very easy to understand, and really helped consolidate the theoretical aspects. I’ve just recently gotten to RNN’s and quite surprised how effective they are. It seems they are used for musical note generations as well.

I was reading through your improvement section, and I don’t quite understand what you mean by this:

‘Train the model on padded sentences rather than random sequences of characters.’

Do you mean to split the text into sentences, and pad each sentence with zeros to match the max length sentence?

71. Prashant Goyal February 12, 2018 at 5:12 am #

As you have used 100 characters to predict the next one, I wanted to know is there any method that I can use to remove this restriction. I want the user to input any character length sentence and generate sentences using it. Thanks.

• Prashant Goyal February 12, 2018 at 5:15 am #

Is there any other way instead of padding the sentences? I really do not want to do this.

• Jason Brownlee February 12, 2018 at 8:33 am #

You can also change the model to operate on one time step at a time and manually reset state.

• Jason Brownlee February 12, 2018 at 8:32 am #

Sure, you have configure it anyway you like. The model will need to tuned for your specific framing.

72. Jo Kanghui February 12, 2018 at 1:14 pm #

Hi, Jason. Thank you for your nice blog.

By the way, when I run this code, I got a ValueError message.

It says “ValueError: You are trying to load a weight file containing 3 layers into a model with 2 layers.”

And I thought that I run this code exactly same as yours, but I got a message like that.

Did I doing something wrong??

• Jo Kanghui February 12, 2018 at 1:48 pm #

Never mind. I did it wrong. The problem is solved.

• Jason Brownlee February 12, 2018 at 2:51 pm #

Sorry to hear that, I have not seen this error. Perhaps the API has changed?

• Kyrylo K June 10, 2018 at 12:14 pm #

How did you fix it?

73. Harry February 22, 2018 at 6:54 am #

Hello Jason,

When you transform input sequences into the form [samples, time steps, features]
using X = numpy.reshape(dataX, (n_patterns, seq_length, 1)) , why is the features equal to 1? I was thinking it’d be equal to the num of chars

• Jason Brownlee February 22, 2018 at 11:22 am #

In that we are providing a sequence of integers.

74. bylo February 28, 2018 at 9:34 pm #

When training on my own data, I notice after around 25 epochs the loss starts to increase. I’ve tried adding a batchnorm layer but it doesn’t do much. Any idea?

75. Mathangi March 5, 2018 at 6:34 am #

Hi Jason

Great article and I am a beginner in Neural Nets. I exactly implemented your code. Ran it for 20 epochs. However while printing the next predicted character it just repeats the same character again and again. I printed out the prediction matrix (before the argmax line) and checked and the prediction matrix is having the same order of values for every prediction. What do you think is going wrong

[[ 1.91119227e-08 1.16459273e-01 1.62738379e-05 2.84793477e-19
3.71392664e-26 1.29606026e-25 8.98123726e-28 1.22408485e-12
4.26335023e-27 1.49551446e-13 6.07275735e-24 7.06751049e-01
1.82651810e-10 4.56609821e-04 7.45972931e-19 1.12589063e-25
8.50236564e-26 4.02262855e-25 4.94167675e-24 2.13961749e-23
1.36306586e-25 2.24255431e-25 1.49555098e-25 7.40160183e-24
4.36129492e-25 2.62904668e-05 6.99173128e-08 1.21143455e-06
4.55503128e-20 1.59668721e-27 2.71070909e-24 9.95265099e-24
1.40039655e-11 2.50412119e-15 8.04228073e-11 5.80116819e-07
1.76183641e-01 7.31929057e-13 4.60703950e-06 1.45222771e-06
1.41010821e-06 8.60679574e-17 5.74646819e-09 3.02204597e-07
8.52235157e-11 7.89179467e-05 1.59914478e-07 3.00487274e-10
1.23463905e-19 5.40685824e-06 5.15879286e-08 5.95685590e-09
1.24319504e-05 5.76499569e-14 1.03425171e-14 1.20372456e-15
7.63502825e-08 6.09846451e-10]]
11
,[[ 1.80287671e-07 2.36380938e-02 2.51635304e-03 4.06853066e-14
1.85676736e-18 1.52850751e-17 6.71450574e-21 5.60100522e-10
1.99784797e-20 1.19577592e-09 7.35863182e-18 9.02304709e-01
4.51892888e-08 1.19447969e-02 2.06239065e-13 9.34988509e-19
2.99471357e-19 3.93370166e-18 9.95959604e-17 1.55780542e-16
3.49815963e-18 8.74736384e-18 1.30014977e-18 8.46453788e-17
5.59418112e-18 3.94404633e-03 1.59909483e-04 5.36000647e-04
8.32731372e-14 2.82467025e-20 2.60687211e-17 3.69919471e-17
2.15683293e-09 3.75411101e-13 1.86202476e-09 2.00714544e-06
5.37619703e-02 6.46848131e-10 4.58007389e-06 1.08297354e-05
1.26086352e-05 1.41199541e-12 3.86868749e-07 4.80629433e-06
1.67563452e-09 4.32701403e-04 3.21613561e-06 2.57872514e-08
1.72215827e-15 7.51974294e-05 1.74515321e-07 1.23122863e-08
6.45807479e-04 5.92429439e-11 6.11113677e-12 5.76062505e-12
8.02828595e-07 6.99018585e-07]]
11

• Jason Brownlee March 6, 2018 at 6:05 am #

The algorithm is stochastic so it may get different results each time it is run.

Perhaps try fitting the model a few times?

76. Federico March 7, 2018 at 1:08 am #

Hi Jason,

I executed the code. Building the model was not a problem. But using the model I hve an error:

ValueError: Dimension 1 in both shapes must be equal, but are 60 and 47 for ‘Assign_5’ (op: ‘Assign’) with input shapes: [256,60], [256,47].

The error is generated at the line of code:

filename = “weights-improvement-20-1.9161.hdf5”

Of course weights-improvement-20-1.9161.hdf5 is my file.

• Jason Brownlee March 7, 2018 at 6:15 am #

Perhaps test making predictions prior to saving and then use the same code after loading so that you know it works.

• Senthil July 24, 2018 at 11:41 pm #

This was caused by the weights.hdf5 file being incompatible with the new data in the repository. I have updated the repo and it should work now.

77. Madhivarman March 9, 2018 at 6:53 pm #

index = numpy.argmax(prediction) only output the maximum value 1. When it converts the index value to char ie.., int_to_char[index] in the dictionary int_to_char the key for 1 is the newline. How should I overcome this?

• Madhivarman March 9, 2018 at 9:18 pm #

working on a smaller dataset cause all this problem.When I trained on larger dataset it actually starts to produce the text. Thanks, @Jason for this article 🙂

• Jason Brownlee March 10, 2018 at 6:24 am #

Perhaps the model requires further tuning on your problem?

78. Andri March 23, 2018 at 1:14 am #

Why didnt you assign batch size(64) same as sequence size(100)?

I still confuse how sequence and batch work when they dont match each other.

• Jason Brownlee March 23, 2018 at 6:10 am #

One batch is comprised of many sequences.

One sequence is one sample or list of time steps.

Does that help?

79. Tom March 29, 2018 at 3:44 pm #

I’ve had some success training a model using words instead of characters. I think it would be interesting to augment each word with synthetic features (parts-of-speech, for instance). But, I can’t wrap my head around how to do this properly? In my mind, I feel that there would need to be a second sparse array with POS-tagging. And perhaps this variable is given a weight of some sort. Does this make sense? Is this possible with the Keras LSTM models?

• Jason Brownlee March 30, 2018 at 6:32 am #

Yes, that makes sense. The input would be a mess and hard to keep straight though.

It might be easier to separate the streams and have a multi-input model instead. Just and idea.
https://machinelearningmastery.com/keras-functional-api-deep-learning/

If you go down this road, I’d love to hear how you go.

• Tom March 30, 2018 at 8:53 am #

Thanks for the response. I’ll work on this and will definitely share the results. I have some other speculative features that I want to experiment with as well.

80. Ray March 31, 2018 at 12:21 pm #

Hi Jason,

If we have 1 million words to predict, shall we still use one hot encoding and softmax in output layer? It might cause memory problem. Is there any way to solve this problem.

Thanks,

Ray

• Jason Brownlee April 1, 2018 at 5:43 am #

Good question.

I have seen some papers that look at splitting up large one hot encoded vectors into multiple pieces. Perhaps try searching on google scholar?

81. Nik April 5, 2018 at 6:02 am #

I am not sure how this code will work?

seq_length = 100
n_patterns = len(dataX)
print “Total Patterns: “, n_patterns
# reshape X to be [samples, time steps, features]
X = numpy.reshape(dataX, (n_patterns, seq_length, 1))

Should this not give an error?

82. jose2007kj April 8, 2018 at 4:33 pm #

sir,thanks for the awesome tutorials,these tutorials are really helpful………
sir what is the procedure for generating a sentance from a set of keywords……..i am having the following data…..

“no_result, sorry, suggest”

from this i need to generate a sentence as following

“Hi…I checked a few options for you, and unfortunately, we do not currently have any trips that meet this criteria. Would you like to book an alternate travel option?”

thanks

• Jason Brownlee April 9, 2018 at 6:07 am #

Good question, I have not worked on this type of problem. Perhaps survey the literature to see what your options are?

83. ML Rookie April 10, 2018 at 6:43 am #

How and where to change the “temperature”, i.e., scale factors?

84. Pierre Lopez April 18, 2018 at 7:40 pm #

Hello,

I am currently working on a project. The idea is to generate the description of a product for example from characteristics and keywords.

My learning base is a set of product descriptions.

I would like to give my model some characteristics, key-words and that it generates me a description from it.

For example :
Characteristics: “Fridge, Bosh, American, Stainless steel, 2 drawers, 531L, 2 vegetable trays”

Generation: “Our experts have selected the BOSH fridge for you: an American stainless steel fridge that keeps all its promises. This product has a total usable volume of 531L, which is very important. With 2 drawers and 2 vegetable trays, it will allow you to store a maximum of fresh produce.”

• Jason Brownlee April 19, 2018 at 6:29 am #

Wow, great problem!

The first step would be to prepare thousands of examples, somehow.

• Chakita January 5, 2021 at 10:55 pm #

Hello!
I am trying to do something similar, did you find a way to do this?
Would it be possible to use a Seq2Seq model for this?

85. Ashok Kumar Harnal May 1, 2018 at 11:05 am #

It is really great that you take time to reply to each question. Something extremely unusual in today’s world.

86. francesca lopez May 27, 2018 at 2:10 am #

Hi Jason!
How come you did not use any validation or test set? Will it be a valid argument to not use any validation or test set as I only want to generate text based on the training data alone? say, i am trying to generate a text based on all harry potter books.. I only want to generate text, words, phrases, sentences based on what my model has learnt and known, since validation and test sets are just there to validate that your trained model can classify and predict what is unknown, it will be of no use already?

• Jason Brownlee May 27, 2018 at 6:48 am #

It is a challenge to test the generative model that is supposed to generate new/different but similar output sequences.

87. Anesh Muthiah May 28, 2018 at 4:09 pm #

Awesome tutorials sir.Can I know what is x in making the prediction??.what do we input during prediction

• Jason Brownlee May 29, 2018 at 6:22 am #

It depends on how the model was defined, e.g. what the model expects as input.

In this tutorial, it expects a seed sequence of 100 words.

88. Dan June 1, 2018 at 9:37 am #

Nice tutorial. With one hot encoding for the input sequences, one can get a loss of 1.22 in less than 10 epochs.

• Jason Brownlee June 1, 2018 at 2:45 pm #

Nice!

• Dan June 1, 2018 at 9:46 pm #

Btw, mainly for the other readers, code to do this is quite simple. just replace lines:

X = numpy.reshape(dataX, (n_patterns, seq_length, 1))
X = X / float(n_vocab)

with:

X = [e for lst in dataX for e in lst]
X = np_utils.to_categorical(X)
enc_length = len(X[0])
X = numpy.reshape(X, (n_patterns, seq_length, enc_length))

Don’t forget to also do this in the text generation part, using enc_length as the second parameter when generating one hot encodings for the seeds.

Also, it’s not compulsory to get seeds from the actual data. You can construct seed/pattern like this:

in_phrase = ‘her name was ‘
in_phrase = [char_to_int[c] for c in in_phrase]
pattern = list(np.ones(100 – len(in_phrase)).astype(int)) + in_phrase

Again, this was/is a really fun and straightforward example to work with. Thanks Jason.

89. Emna June 7, 2018 at 7:16 pm #

Thank you for this nice tutorial. I didn’t really get why you used one hot encoding only for the output character? why for example you didn’t use integer encoding for the output pattern, it will calculate also the output probability for it if it was encoded as the input.

• Jason Brownlee June 8, 2018 at 6:08 am #

An integer encoding was used for the inputs and passed directly to the LSTM.

90. Emna June 7, 2018 at 8:03 pm #

I have another question, Can we use here also word2vec instead of converting the characters to integers and then scaling. Or it won’t be character to character model anymore ?

• Jason Brownlee June 8, 2018 at 6:10 am #

Perhaps, I have not seen embedding models for chars, but I bet it has been tried.

91. Arnab Dhar June 20, 2018 at 3:00 am #

Hello Jason,

I have adapted your code in kaggle.com to try and fit project gutenberg txt files on Shakespeare’s plays mentioning you and this website.

92. Priyanshu Kumar June 28, 2018 at 4:47 am #

Hello Sir, thank you for the post!
I have created this model and trained it on a different data set. The results contains sequences repeating infinitely in a loop. Can you please share some insights?

• Jason Brownlee June 28, 2018 at 6:26 am #

Perhaps the model is over fit?

Perhaps try fitting the model again?

93. Ferin July 12, 2018 at 3:39 pm #

When I tried running the final complete code it shows an error saying im trying to load a weight file containing 2 layers into a model with 3 layers.

• Jason Brownlee July 13, 2018 at 7:31 am #

I have not seen this, are you sure you copied all of the code without modification?

Are you sure you have the lates version of the libraries installed?

94. Raviteja July 15, 2018 at 2:33 am #

int to char is trowing eerror
i have encoded file with utf 8

95. Raviteja July 15, 2018 at 5:12 am #

result = int_to_char[index] i am getting error
here when rinning final code in

• Jason Brownlee July 15, 2018 at 6:19 am #

Are you able to confirm that all of your libraries are up to date and that you copied all of the code from the tutorial?

• ravi July 15, 2018 at 9:34 am #

Yes jason but errror at result line which converts to char 3425

96. herawati July 18, 2018 at 12:14 am #

97. Adam July 20, 2018 at 9:23 pm #

How did you settle on a 256×256 hidden layer?

I ask because I’m interested in not paring out caps, and the vocab in what I’m learning on has expanded to 132 characters. In doing so I’ve gotten the loss down to 1.3, and the generative text is still producing a *LOT* of typos.

If I added more neurons to the LSTM layers, could the bot improve? How many neurons?

Or would it be more beneficial to allow the network to train to an even lower loss on the current 256×256 network?

98. Ahmed Ibrahim July 22, 2018 at 1:39 am #

Hi Jason,

As you mentioned, we can also experiment with other ASCII data, such as computer source code.

So I need to use this post to create a program repair model.
The input will be a buggy code and the output will be the fixed code.

I have some vulnerable/buggy examples and their fixes but I don’t know how to generate a dataset and make a train for that.

Thanks

• Jason Brownlee July 22, 2018 at 6:26 am #

Sounds like a cool project.

Sorry, i don’t have examples of repairing bugging code with LSTMs, perhaps in the future.

99. Fawaz July 22, 2018 at 4:23 pm #

Hi. Thanks for this post. I have one question. How come you aren’t providing the output labels of every timestep? For example when the input is ‘HelloWorl’, then the output is ‘elloWorld’, if we are using 9 timesteps. For your example, there is only one letter for each sample. How are you going to provide the corresponding outputs for the timesteps then during training?
Thanks

• Jason Brownlee July 23, 2018 at 6:08 am #

I’m not sure I follow, sorry?

100. Peter Wendel July 23, 2018 at 5:37 am #

hey jason, check out a blog post i made that leverages some of you methodology!
this post is awesome!
http://overslant.com/2018/07/22/deep-nba-nicknames/

101. Didier G. July 25, 2018 at 2:20 am #

Hi Jason,
Thank you for the nice article. It made many things easy for me… haha.
I have one question though. I don’t quite understand the necessity of line 65 in the full code listing of the “Generating text with and LSTM Network” section:

seq_in = [int_to_char[value] for value in pattern]

seq_in is set but doesn’t seemed to be used.

• Jason Brownlee July 25, 2018 at 6:21 am #

If it is not used, you can ignore it, delete the line.

102. verdy August 3, 2018 at 4:36 pm #

HI…Jason

Please give me information a baut specification minimum ( ram, processor, etc) a laptop for running 3D-Unet-Pytorch for classification images 2D/3D.and can i try MINST data set for 3D images.

Thanks.

Regards,,

Verdy

• Jason Brownlee August 4, 2018 at 6:00 am #

You can train on your CPU or use AWS if you need to access GPUs.

103. owais August 13, 2018 at 4:53 am #

With same as your code I trained model but during testings I got this…

ValueError: You are trying to load a weight file containing 1 layers into a model with 2 layers.

Any idea what’s the problem?

104. Curious August 30, 2018 at 2:10 pm #

shoudnt the input sequnce and lstm size be tha same, eg. here since sample input is 100 characters lstm should be 100 as well instead of 256

• Jason Brownlee August 30, 2018 at 4:52 pm #

The number of nodes in the first hidden layer is unrelated to the number of units (defined by input_shape).

105. kaiser September 11, 2018 at 9:48 pm #

how did you decide the number of hidden units? ,can we take 128 instead of 256

106. Kwaku Elikplim September 20, 2018 at 8:57 am #

Hello Jason, great tutorial, as always!!
I tried running this however on Kaggle, the training went well, got a good saved weights. However, during generation, I ran into the error and I can’t seem to identify what is causing it:

“—————————————————————————
KeyError Traceback (most recent call last)
in ()
79 prediction = model.predict(X, verbose=0)
80 index = numpy.argmax(prediction)
—> 81 result = int_to_char[index]
82 seq_in = [int_to_char[value] for value in pattern]
83 sys.stdout.write(result)

KeyError: 2463279”

107. Uday October 9, 2018 at 8:54 pm #

Hi Jason,

Great Post!

Just curious to know the possibilities of implementing Sequence mining use case here.

Can we use Generative LSTM Networks for sequence mining? This is to suppress the events.

I have a dataset with 2 million sequence of events, where we have root events and child events. We have to suppress all the child events based on root event presence. This pattern needs to captured by Generative LSTM model which is trained on 70% of the whole data. And then apply this trained model on test data (30%) to perform event suppression to dedue the whole number of events.

Thanks,
Uday Vakalapudi

• Jason Brownlee October 10, 2018 at 6:08 am #

Sorry, I don’t know what “sequence mining” is?

108. Mario October 14, 2018 at 6:46 pm #

Great post. It inspire me to think about the application of the solution to my problem. So, let see what do you think about LSTM capabilities to predict next value in the array like this: [5.3, 2.2, 3.1, 8.33, 2.32, 2.01, 4, 12.2, 12.2, 4, 4, 3, 2.30, 13.30…]? It is like the sentences if you look from this perspective:
– first three numbers are connected like this: first number is the sum of second and third;
– forth number isn’t connected with the previous and it represented the sum of the next three numbers;

General conclusion about array: there is a groups of numbers they are connected with the mathematical operations of subtraction and additions but there is no connection between groups except that they represented the journal entries in a bookeeping.

• Jason Brownlee October 15, 2018 at 7:26 am #

It may be possible, perhaps try it and see.

109. Poppie360 November 2, 2018 at 3:55 am #

I got this error when using your code, any help or advice you could give me?
https://i.gyazo.com/69b94f1f42990146b27050dd2459a3f3.png

• Jason Brownlee November 2, 2018 at 5:57 am #

You must change the filename to the file in the code that you saved.

• Poppie360 November 2, 2018 at 9:24 am #

I just might not have seen the part, but could you please tell me what file that would be? I didn’t see a part about saving a file.

• Jason Brownlee November 2, 2018 at 2:49 pm #

• Poppie360 November 2, 2018 at 10:37 pm #

Thank you, i accidentally missed a part, have a good day and thank you for the help

• Jason Brownlee November 3, 2018 at 7:06 am #

No problem.

• vera March 20, 2019 at 12:23 am #

I am new here,can i use this model to train a text time-based collection of documents and then predict the documents that will be generated at a future time point.

• Jason Brownlee March 20, 2019 at 8:32 am #

Perhaps, you may have to experiment a little to discover a suitable model.

• vera March 21, 2019 at 1:09 am #

Can you recommend some papers or documents that have done this to me ? Thank you Jason .

• Jason Brownlee March 21, 2019 at 8:17 am #

• vera March 21, 2019 at 1:58 am #

can you recommend some papers about this that use LSTM and LDA to predict the Technology topic Trend, I’m stuck here.Thanks very much

110. Palak November 13, 2018 at 6:55 pm #

Hi Jason, I follow your posts and they are absolutely great. But I am trying to generate text using tensorFlow.
I have trained my model using tensorflow’s original RNN code; https://www.tensorflow.org/tutorials/sequences/recurrent#tutorial_files; but not sure how to predict and test my model. Looking forward to hear from you.
Thanks for the great work!

• Jason Brownlee November 14, 2018 at 7:28 am #

Sorry, I don’t have TensorFlow tutorials, I focus on Keras that runs on top of TensorFlow.

111. Beginner November 13, 2018 at 6:58 pm #

Thanks!

112. savaş türkoğlu December 23, 2018 at 3:57 am #

Hi Jason thank you for this great tutorial.

• Jason Brownlee December 23, 2018 at 6:08 am #

You’re welcome, I’m glad it helped.

• Ritz January 5, 2019 at 10:27 pm #

Hi,
In this tutorial, if you keep training until you get really good accuracy, say 99%, isn’t that just memorizing the data? The end goal is to generate something new right?

• Jason Brownlee January 6, 2019 at 10:18 am #

Yes, the idea is to have a dataset that is large enough or a model that is regularized enough that it cannot be memorized.

113. Nathan January 8, 2019 at 10:18 pm #

Hi Jason,

I have seen your tutorial and tried it.
I am now using a different project, where we want to launch an LSTM network in production.

In other words, we have a Time Series Prediction network, and we want to place it on AWS or Azure. We have also seen Tensorflow Serving as a way of putting these Networks online.

The only thing i’m afraid of is the performance of LSTM.
Let me elaborate:
When training an LSTM network, the Long- and Short-term memory is crucial. The weights get adapted as well, so they are important.
If we save the model – like you did in the tutorial – will we still get good results if we try to run a test-seed through it?

Some seed it has never seen before, like some kind of user-input? I’m afraid that it will not get great results, as the Hidden states are lost now, and the seed will make for new hidden states, that are completely different.
Or am I totally wrong with this?

• Jason Brownlee January 9, 2019 at 8:45 am #

Perhaps try using an ensemble of final models to reduce the variance?

Perhaps design tests for the system before deploying it into production (e.g. good engineering practices).

114. Mar January 19, 2019 at 8:18 am #

Consider having a transform that converts a word (or a few words) to a binary vector, multi-hot.
Now, I want to convert back the binary vector to the original word (or words).
Can this be done using RNN/LSTM?

Of course, I can train the model using a train set (given the words and their corresponding binary vector), but, then, test it with a predicted binary vector, hopefully, to predict the correct words.

• Jason Brownlee January 20, 2019 at 5:37 am #

Yes, calculate the argmax() of each vector, then map the integer to the word in your vocab – a reverse lookup.

I believe I have examples in many tutorials of this.

115. Jason January 28, 2019 at 7:16 am #

Hi Jason ,

Is it possible to generate sequence using return_sequence=False ?
Meaning that only the last timestep is being predicted in train – how would a generation look like ?

Thanks

• Jason Brownlee January 28, 2019 at 11:43 am #

I’m not sure I understand your question, sorry, can you elaborate please?

You can frame any sequence prediction problem you wish and test whether an LSTM performs well or not.

• Jason January 29, 2019 at 6:33 am #

In your example you use this line :

As far as I understand , the return_sequences=True means that there will be an output in every item in the sequence (many-to-many) and Tx=Ty. and if return_sequences=False there will be only one output to the sequence (of the last) and only it will be accounted for in the Loss function.

So , can return_sequences=False be used in a generation-task such as this ?

• Jason Brownlee January 29, 2019 at 11:38 am #

Yes, but the results will be poor when making a prediction with the output of the LSTM layer directly. It is better to use one or more fully connected layers to interpret the output of the LSTM.

116. Krish February 14, 2019 at 11:28 am #

Hi jason,
Thank you for the informative tutorial.
I am on python 2.7, tensorflow 1.12, and keras 2.2.3
I am getting the following error:

maximum_iterations=input_length)
TypeError: while_loop() got an unexpected keyword argument ‘maximum_iterations’
ERROR:tensorflow:==================================
Object was never used (type ):

• Jason Brownlee February 14, 2019 at 2:17 pm #

Sorry, I have not seen this error. Perhaps try searching/posting on stackoverflow?

117. John February 15, 2019 at 5:47 am #

Hi Jason,
Thank you for the tutorial.
I am repurposing this model to generate SMILES strings, which are character representations of drugs and molecules:
They look like this:
Nc1nc(Nc2ccc(cc2)S(N)(=O)=O)nn1C(=O)c1c(F)cccc1F
CO[[email protected]]1[[email protected]]
I have trained the model and saved the weights.
How do I now get the model to generate the SMILES strings with variable lengths.
Thanks,
John

• Jason Brownlee February 15, 2019 at 8:18 am #

Perhaps train the model on zero-padded but variable length content inputs that have a “end of string” char (e.g. like a period). Then generate new sequences and stop when that char is encountered.

Does that help?

118. John February 15, 2019 at 10:03 am #

Hi Jason,
Thank you for the advice, but what if it is important for the model to understand the end of the sequence or there is a specific pattern at the end.

• Jason Brownlee February 15, 2019 at 2:21 pm #

Sorry, I don’t follow, what is the concern exactly?

119. Koushik J February 17, 2019 at 6:12 pm #

Hey Jason Brownlee.
Thank you for the wonderful explanation about the text generation and also providing the code with point to point explanation.

• Jason Brownlee February 18, 2019 at 6:28 am #

You’re welcome, I’m glad it helped.

120. Samuel Muiruri February 19, 2019 at 8:23 pm #

I’m running this model (the simple LSTM) but with what you’d call a huge dataset. A file of about 350 mb and this much info

[email protected]:~/rap-generator\$ python new_model.py
Using TensorFlow backend.
(‘Total Characters: ‘, 307478020)
(‘Total Vocab: ‘, 177)
Killed

which btw I’m running on a google cloud instance with 120 GB of RAM but it exhausts it all.

Here’s the file so far https://bitbucket.org/muiruri_samuel/rap-generator/src/master/new_model.py

I’ve contemplated just training on a file with 1m lines of text and that’s become 10 smaller files of 30 mbs but I wonder is it possible to know how much the main file would need in RAM (lyrics.txt) and can I get it to work in a batch method either using the smaller files and a kind of fit_generator method.

• Jason Brownlee February 20, 2019 at 8:02 am #

You can estimate the RAM based on the number of chars and the choice of 8-bit or 16-bit encoding.

Perhaps progressive loading with a custom data generator would be the way to go?

121. Sanjeev March 19, 2019 at 4:17 pm #

Hi Jason,

This is a wonderful post, thanks for sharing. I have a different problem to solve and was wondering if it can be solved using your solution.

I have a corpus of numerical data (structured) and its corresponding article (readable text). i.e. A simple table with numbers in it and a paragraph explaining the table. How can this dataset be used to train the model and then used to generate paragraphs based on any input of the table form?

Sanjeev

• Jason Brownlee March 20, 2019 at 8:24 am #

Wow, great problem.

Try a model that might use an MLP, CNN or LSTM to read in the numbers, and then a decoder to output text – maybe one output for the table and one for the text.

122. Lokesh April 2, 2019 at 10:21 pm #

Hello Jason,

Thank you for such an amazing post. I did a complete tensorflow version of the same.

It would be great if you can write some blogs on BERT and GPT.

Highly appreciate the work you’re doing for the AI community.

Regards,
LK

123. Haru April 16, 2019 at 5:12 am #

Hi Jason, Thanks for the tutorial.

Could you please shed me some light on how I should approach for the following problem:

I have a set of words. The word(s) might have spelling error(s).
From the set of words, I would like to generate a sentence.
The set of words might ( in most of the cases ) be in the order of the sentence formation, BUT there might be missing words also.

e.g: A dog is running behind a car. (ORIGINAL sentence)

word_set={‘a’, ‘dot’, ‘runni’, ‘ehin’, ‘ar’}

I am working on multiple NON English languages, where tools such as POS tagger, etc are NOT available.
For one language, I have some pretty good amount of corpus. And for one, very less amount of corpus.
The sentence that I have to generate using the set of words may or may not be present in the corpus.

(Any steps or tutorial guide will be helpful).

• Jason Brownlee April 16, 2019 at 6:55 am #

I think some development and prototyping might be required – there’s no step-by-step tutorial for this.

Perhaps you can get some ideas from related text-correction papers?

• Haru April 16, 2019 at 7:31 pm #

I was think of spelling correction followed by text sequence generation. Anyway, I will look into it. Thanks!

124. Alejandro Oñate Latorre April 16, 2019 at 7:30 am #

Hello, I would like to know your opinion on why it is better to generate text predicting letter by letter and not word by word.

I’m trying to make a text generator in Spanish, with the little prince’s book, the results letter by letter and word by word are different but none I like.

Letter by letter I get sentences with more global sense, but with incorrect letters. But word by word I get incoherent sentences.

BR!

• Jason Brownlee April 16, 2019 at 2:17 pm #

Great question.

I don’t know, perhaps you can design some experiments to help tease out the cause and effect.

Perhaps it has to do with the limited cardinality if the input/output (letters vs words).

125. Chirag prajapati April 29, 2019 at 3:57 pm #

hii great blog thanks for it

I have also implemented by referencing your blog. I had used shakespeare poem.
with 30 epochs 64 batch size I got “weights-improvement-30-1.4482.hdf5”
error loss 1.4482

I have one request can you write a blog on Recommendation system with RNN LSTM in keras

Once again thank you

• Jason Brownlee April 30, 2019 at 6:47 am #

Well done!

Thanks for the suggestion.

126. Rajesh May 2, 2019 at 1:16 pm #

Hi Jason,
In all the LSTM text generative models I found online, there’s a text file and they’re predicting the next sequence. In my case, I’ve an input and output column in a csv file. For a particular input, I want the model to generate corresponding output. How do I approach this? and what are the pre-processing steps involved?

• Jason Brownlee May 2, 2019 at 2:03 pm #

You can load the data into memory first, then work with it any way you wish.

• Rajesh May 8, 2019 at 8:48 pm #

I really didn’t get that. What I did was, I converted my input and output column to two individual text files, did character mapping for both of them individually, then taking X and y according to a sequence length. But couldn’t fit the model. I’m getting no.of input samples doesn’t match output samples.
Could you elaborate the steps that have to be done. Again, I have an input text column and dependent output text column. Based on an input sequence, the model has to generate output column’s text

• Jason Brownlee May 9, 2019 at 6:41 am #

127. Jignesh Waghela June 5, 2019 at 8:41 pm #

Hi Jason. Nice post on text generation.

I have few queries regarding my ongoing project:

How do I parse/process the structured data(dataframe format) for text generation, where-in I can the output text in a form of sentences. Eg.: Lets say I have Financial data in csv/excel format and once I load it in NLP, I should get the insights of data in the form of narratives/sentences.

I know this is more of the NLG task, but I am not able to set a pipeline right from parsing the data to generating narratives.

• Jason Brownlee June 6, 2019 at 6:24 am #

I’m not sure off hand, experimentation will be required.

128. Bryce Beckwith June 24, 2019 at 7:30 am #

Thanks Jason this is an awesome post! I modified my code to treat the characters as actual words as my base dataset is extremely small ~200 sentences. I plan on posting my example on Github. Is there a specific way you like to be cited?

Thanks!

129. Peter June 25, 2019 at 7:37 pm #

I am interested to try this example out. You mentioned that with GPU it takes about 700 seconds per epoch for the 2-hidden-layered-LSTM. How many seconds (or hours) would it take for Intel i5 CPU, if you could give some raw estimate?

• Jason Brownlee June 26, 2019 at 6:38 am #

I don’t know, sorry. Perhaps test it and find out?

130. morgan July 4, 2019 at 7:37 am #

hey jason,
why are you using the seed sent (to test the model)extracted from the exact same text(corpus) that you have used to train/fit your model? isn’t it obvious that the model you have trained on the corpus will generate the same output during testing if it sees the same sequence of char in its training data!!! the real test should be using a sentence different from the corpus OR it does not matter that much?!?!?

131. Ruhan siddiqui July 4, 2019 at 5:53 pm #

thanks for this code sir
due to your tutorial i have learned many think for neural network

• Jason Brownlee July 5, 2019 at 7:51 am #

You’re welcome. I’m happy to hear that.

132. Kakoli July 25, 2019 at 2:09 pm #

Thanks for the easy-to-understand post.
1 question though : One improvement suggested was add dropout to the visible input layer.
Does it mean the parameter dropout to LSTM function?

• Jason Brownlee July 25, 2019 at 2:13 pm #

Before the first LSTM layer, e.g. dropout as the first hidden layer layer.

• Kakoli July 26, 2019 at 6:05 am #

Thanks Jason for your quick response. But I did not see any improvement on adding Dropout before LSTM layer.

In your code, input is a sliding window of 100 chars and output is 101st char.
Now for training the model on padded sentences, I have converted the input into padded sentences of 100 words. But how do I model the input and output? For each sentence of 100 words, the output is the101st word. Any better way?

• Jason Brownlee July 26, 2019 at 8:35 am #

You might have to get creative and explore a range of ideas to see what works well/best.

• Kakoli July 26, 2019 at 12:30 pm #

The trouble here is this : to explore even 1 idea, takes min 50 epochs to see its proof, with each epoch ~12mins. So was asking if you have any experience.

Thanks again for the great post and your prompt responses.

• Jason Brownlee July 26, 2019 at 2:19 pm #

Opinions are not helpful because models and data can vary so widely, I would encourage you to experiment.

Perhaps you can reduce the size of the dataset or model to increase the rate of testing ideas?

133. ashraf July 29, 2019 at 4:59 am #

sir in spyder it is showing that

Unable to open file (unable to open file: name = ‘weights-improvement-19-1.9435.hdf5’, errno = 2, error message = ‘No such file or directory’, flags = 0, o_flags = 0)

how to solve this problem

134. Agam Singh August 6, 2019 at 5:59 pm #

@jasonBrownlee can we ask the model to generate a sentence based on few input keywords given, instead of random seeded sentence it generated

• Jason Brownlee August 7, 2019 at 7:43 am #

That would be a different problem, e.g. generate a sequence from keywords.

It would be like the reverse of text summarization.

135. priyanshu Khullar August 11, 2019 at 10:53 pm #

Jason Brother it is prinitng empty letters Any suggestions ?

• Jason Brownlee August 12, 2019 at 6:37 am #

Perhaps try fitting the model again?

136. Simon August 28, 2019 at 1:45 am #

Hi, very nice website!

I am trying to generate swim practices programs from learning all the swim practice that I did. I think this model would apply. Do you think It could be adapted? I have a basic setup now that is giving me some results, but I would like to add more data to each exercise, like time. Do you think I am in the right path ?

• Jason Brownlee August 28, 2019 at 6:40 am #

I don’t know sorry. Perhaps experiment and see what you can achieve?

137. Muhammad Usgan August 28, 2019 at 2:00 pm #

Sir, how do I determine the correct batch size?

• Jason Brownlee August 28, 2019 at 3:01 pm #

Great question, I recommend testing a range of different batch sizes to see what works well on your specific problem.

138. Augusto August 29, 2019 at 10:23 pm #

Hi Jason,

I have tried to run the 2nd exercise using 50 epochs, but in my PC it simple does not finish, it crash after 2 hrs. Even I tried in a AWS machine (ml.m4.xlarge) but it hangs also.