Last Updated on August 14, 2019
Recurrent neural networks, or RNNs, are a type of artificial neural network that add additional weights to the network to create cycles in the network graph in an effort to maintain an internal state.
The promise of adding state to neural networks is that they will be able to explicitly learn and exploit context in sequence prediction problems, such as problems with an order or temporal component.
In this post, you are going take a tour of recurrent neural networks used for deep learning.
After reading this post, you will know:
- How top recurrent neural networks used for deep learning work, such as LSTMs, GRUs, and NTMs.
- How top RNNs relate to the broader study of recurrence in artificial neural networks.
- How research in RNNs has led to state-of-the-art performance on a range of challenging problems.
Kick-start your project with my new book Long Short-Term Memory Networks With Python, including step-by-step tutorials and the Python source code files for all examples.
Note, we’re not going to cover every possible recurrent neural network. Instead, we will focus on recurrent neural networks used for deep learning (LSTMs, GRUs and NTMs) and the context needed to understand them.
Let’s get started.
We will start off by setting the scene for the field of recurrent neural networks.
Next, we will take a closer look at LSTMs, GRUs, and NTM used for deep learning.
We will then spend some time on advanced topics related to using RNNs for deep learning.
- Recurrent Neural Networks
- Fully Recurrent Networks
- Recursive Neural Networks
- Neural History Compressor
- Long Short-Term Memory Networks
- Gated Recurrent Unit Neural Networks
- Neural Turing Machines
Recurrent Neural Networks
Let’s set the scene.
Popular belief suggests that recurrence imparts a memory to the network topology.
A better way to consider this is the training set contains examples with a set of inputs for the current training example. This is “conventional, e.g. a traditional multilayered Perceptron.
X(i) -> y(i)
But the training example is supplemented with a set of inputs from the previous example. This is “unconventional,” e.g. a recurrent neural network.
[X(i-1), X(i)] -> y(i)
As with all feed forward network paradigms, the issues are how to connect the input layer to the output layer, include feedback activations, and then train the construct to converge.
Let’s now take a tour of the different types of recurrent neural networks, starting with very simple conceptions.
Fully Recurrent Networks
The layered topology of a multilayer Perceptron is preserved, but every element has a weighted connection to every other element in the architecture and has a single feedback connection to itself.
Not all connections are trained and the extreme non-linearity of the error derivatives means conventional Backpropagation will not work, and so Backpropagation Through Time approaches or Stochastic Gradient Descent is employed.
Also, see Bill Wilson’s Tensor Product Networks (1991).
Recursive Neural Networks
Recurrent neural networks are linear architectural variant of recursive networks.
Recursion promotes branching in hierarchical feature spaces and the resulting network architecture mimics this as training proceeds.
Training is achieved with Gradient Descent by sub-gradient methods.
This is described in detail in R. Socher, et al., Parsing Natural Scenes and Natural Language with Recursive Neural Networks, 2011.
Neural History Compressor
Schmidhuber reported a very deep learner, first in 1991, that was able to perform credit assignment over hundreds of neural layers by unsupervised pre-training for a hierarchy of RNNs.
Each RNN is trained unsupervised to predict the next input. Then only inputs generating an error are fed forward, conveying new information to the next RNN in the hierarchy, which then processes at a slower, self-organizing time scale.
It was shown that no information is lost, just compressed. The RNN stack is a “deep generative model” of the data. The data can be reconstructed from the compressed form.
See J. Schmidhuber, et al., Deep Learning in Neural Networks: An Overview, 2014.
Backpropagation fails as the calculation of extremity of non-linear derivatives increases as the error is propagated backwards through large topologies, making credit assignment difficult, if not impossible.
Need help with LSTMs for Sequence Prediction?
Take my free 7-day email course and discover 6 different LSTM architectures (with code).
Click to sign-up and also get a free PDF Ebook version of the course.
Long Short-Term Memory Networks
With conventional Back-Propagation Through Time (BPTT) or Real Time Recurrent Learning (RTTL), error signals flowing backward in time tend to either explode or vanish.
The temporal evolution of the back-propagated error exponentially depends on the size of the weights. Weight explosion may lead to oscillating weights, while in vanishing causes learning to bridge long time lags and takes a prohibitive amount of time, or does not work at all.
- LSTM is a novel recurrent network architecture training with an appropriate gradient-based learning algorithm.
- LSTM is designed to overcome error back-flow problems. It can learn to bridge time intervals in excess of 1000 steps.
- This true in presence of noisy, incompressible input sequences, without loss of short time lag capabilities.
Error back-flow problems are overcome by an efficient, gradient-based algorithm for an architecture enforcing constant (thus neither exploding nor vanishing) error flow through internal states of special units. These units reduce the effects of the “Input Weight Conflict” and the “Output Weight Conflict.”
The Input Weight Conflict: Provided the input is non-zero, the same incoming weight has to be used for both storing certain inputs and ignoring others, then will often receive conflicting weight update signals.
These signals will attempt to make the weight participate in storing the input and protecting the input. This conflict makes learning difficult and calls for a more context-sensitive mechanism for controlling “write operations” through input weights.
The Output Weight Conflict: As long as the output of a unit is non-zero, the weight on the output connection from this unit will attract conflicting weight update signals generated during sequence processing.
These signals will attempt to make the outgoing weight participate in accessing the information stored in the processing unit and, at different times, protect the subsequent unit from being perturbed by the output of the unit being fed forward.
These conflicts are not specific to long-term lags and can equally impinge on short-term lags. Of note though is that as lag increases, stored information must be protected from perturbation, especially in the advanced stages of learning.
Network Architecture: Different types of units may convey useful information about the current state of the network. For instance, an input gate (output gate) may use inputs from other memory cells to decide whether to store (access) certain information in its memory cell.
Memory cells contain gates. Gates are specific to the connection they mediate. Input gates work to remedy the Input Weight Conflict while Output Gates work to eliminate the Output Weight Conflict.
Gates: Specifically, to alleviate the input and output weight conflicts and perturbations, a multiplicative input gate unit is introduced to protect the memory contents stored from perturbation by irrelevant inputs and a multiplicative output gate unit protects other units from perturbation by currently irrelevant memory contents stored.
Connectivity in LSTM is complicated compared to the multilayer Perceptron because of the diversity of processing elements and the inclusion of feedback connections.
Memory cell blocks: Memory cells sharing the same input gate and the same output gate form a structure called a “memory cell block”.
Memory cell blocks facilitate information storage; as with conventional neural nets, it is not so easy to code a distributed input within a single cell. A memory cell block of size 1 is just a simple memory cell.
Learning: A variant of Real Time Recurrent Learning (RTRL) that takes into account the altered, multiplicative dynamics caused by input and output gates is used to ensure non-decaying error back propagated through internal states of memory cells errors arriving at “memory cell net inputs” do not get propagated back further in time.
Guessing: This stochastic approach can outperform many term lag algorithms. It has been established that many long-time lag tasks used in previous work can be solved more quickly by simple random weight guessing than by the proposed algorithms.
See S. Hochreiter and J. Schmidhuber, Long-Short Term Memory, 1997.
The most interesting application of LSTM Recurrent Neural Networks has been the work done with language processing. See the work of Gers for a comprehensive description.
- F. Gers and J. Schmidhuber, LSTM Recurrent Networks Learn Simple Context Free and Context Sensitive Languages, 2001.
- F. Gers, Long Short-Term Memory in Recurrent Neural Networks, Ph.D. Thesis, 2001.
The efficient, truncated version of LSTM will not easily solve problems similar to “strongly delayed XOR.”
Each memory cell block needs an input gate and an output gate. Not necessary in other recurrent approaches.
Constant error flow through “Constant Error Carrousels” inside memory cells produces the same effect as a conventional feed-forward architecture being presented with the entire input string at once.
LSTM is as flawed with the concept of “regency” as other feed-forward approaches. Additional counting mechanisms may be required if fine-precision counting time steps is needed.
The algorithms ability to bridge long time lags is the result of constant error Backpropagation in the architecture’s memory cells.
LSTM can approximate noisy problem domains, distributed representations, and continuous values.
LSTM generalizes well over problem domains considered. This is important given some tasks are intractable for already established recurrent networks.
Fine tuning of network parameters over the problem domains appears to be unnecessary.
In terms of update complexity per weight and time steps, LSTM is essentially equivalent to BPTT.
LSTMs are showing to be powerful, achieving state-of-the-art results in domains like machine translation.
Gated Recurrent Unit Neural Networks
Gated Recurrent Neural Networks have been successfully applied to sequential or temporal data.
Most suitable for speech recognition, natural language processing, and machine translation, together with LSTM they have performed well with long sequence problem domains.
Gating was considered in the LSTM topic and involves a gating network generating signals that act to control how the present input and previous memory work to update the current activation, and thereby the current network state.
Gates are themselves weighted and are selectively updated according to an algorithm, throughout the learning phase.
Gate networks introduce added computational expense in the form of increased complexity, and therefore added parameterization.
The LSTM RNN architecture uses the computation of the simple RNN as an intermediate candidate for the internal memory cell (state). The Gated Recurrent Unit (GRU) RNN reduces the gating signals to two from the LSTM RNN model. The two gates are called an update gate and a reset gate.
The gating mechanism in the GRU (and LSTM) RNN is a replica of the simple RNN in terms of parameterization. The weights corresponding to these gates are also updated using BPTT stochastic gradient descent as it seeks to minimize a cost function.
Each parameter update will involve information pertaining to the state of the overall network. This can have detrimental effects.
The concept of gating is explored further and extended with three new variant gating mechanisms.
The three gating variants that have been considered are, GRU1 where each gate is computed using only the previous hidden state and the bias; GRU2, where each gate is computed using only the previous hidden state; and GRU3, where each gate is computed using only the bias. A significant reduction in parameters is observed with GRU3 yielding the smallest number.
The three variants and the GRU RNN were benchmarked using data from the MNIST Database of handwritten digits and the IMDB movie review dataset.
Two sequences lengths were generated from the MNIST dataset and one was generated from the IMDB dataset.
The main driving signal of the gates appears to be the (recurrent) state as it contains essential information about other signals.
The use of the stochastic gradient descent implicitly carries information about the network state. This may explain the relative success in using the bias alone in the gate signals as its adaptive update carries information about the state of the network.
Gated variants explore the mechanisms of gating with limited evaluation of topologies.
For more information see:
- R. Dey and F. M. Salem, Gate-Variants of Gated Recurrent Unit (GRU) Neural Networks, 2017.
- J. Chung, et al., Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling, 2014.
Neural Turing Machines
Neural Turing Machines extend the capabilities of neural networks by coupling them to external memory resources, which they can interact with through attention processes.
The combined system is analogous to a Turing Machine or Von Neumann architecture, but is differentiable end-to-end, allowing it to be efficiently trained with gradient descent.
Preliminary results demonstrate that Neural Turing Machines can infer simple algorithms, such as copying, sorting, and associative recall from input and output examples.
RNNs stand out from other machine learning methods for their ability to learn and carry out complicated transformations of data over extended periods of time. Moreover, it is known that RNNs are Turing-Complete and therefore have the capacity to simulate arbitrary procedures, if properly wired.
The capabilities of standard RNNs are extended to simplify the solution of algorithmic tasks. This enrichment is primarily via a large, addressable memory, so, by analogy to Turing’s enrichment of finite-state machines by an infinite memory tape, and so dubbed “Neural Turing Machine” (NTM).
Unlike a Turing machine, an NTM is a differentiable computer that can be trained by gradient descent, yielding a practical mechanism for learning programs.
Crucially, every component of the architecture is differentiable, making it straightforward to train with gradient descent. This was achieved this by defining ‘blurry’ read-and-write operations that interact to a greater or lesser degree with all the elements in memory (rather than addressing a single element, as in a normal Turing machine or digital computer).
For more information see:
- A. Graves, et al., Neural Turing Machines, 2014.
- R. Greve, et al., Evolving Neural Turing Machines for Reward-based Learning, 2016.
The copy task tests whether NTM can store and recall a long sequence of arbitrary information. The network is presented with an input sequence of random binary vectors followed by a delimiter flag.
The networks were trained to copy sequences of eight-bit random vectors where the sequence lengths were randomized between 1 and 20. The target sequence was simply a copy of the input sequence (without the delimiter flag).
Repeat copy task extends copy by requiring the network to output the copied sequence a specified number of times and then emit an end-of-sequence marker. The main motivation was to see if the NTM could learn a simple nested function.
The network receives random-length sequences of random binary vectors, followed by a scalar value indicating the desired number of copies, which appears on a separate input channel.
Associative recall tasks involve organizing data arising from “indirection”, that is when one data item points to another. A list of items is constructed so that querying with one of the items demands that the network returns the subsequent item.
A sequence of binary vectors that is bounded on the left and right by delimiter symbols is defined. After several items have been propagated to the network, the network is queried by showing a random item, and seeing if the network can produce the next item.
Dynamic N-Grams task tests if the NTM can adapt quickly to new predictive distributions by using memory as a re-writable table that it could use to keep count of transition statistics, thereby emulating a conventional N-Gram model.
Consider the set of all possible 6-Gram distributions over binary sequences. Each 6-Gram distribution can be expressed as a table of 32 numbers, specifying the probability that the next bit will be one, given all possible length five binary histories. A particular training sequence was generated by drawing 200 successive bits using the current lookup table. The network observes the sequence one bit at a time and is then asked to predict the next bit.
Priority sort task tests the NTM’s ability to sort. A sequence of random binary vectors is input to the network along with a scalar priority rating for each vector. The priority is drawn uniformly from the range [-1, 1]. The target sequence contains the binary vectors sorted according to their priorities.
NTMs have feed-forward architectures to LSTMs as one of their components.
In this post, you discovered recurrent neural networks for deep learning.
Specifically, you learned:
- How top recurrent neural networks used for deep learning work, such as LSTMs, GRUs, and NTMs.
- How top RNNs relate to the broader study of recurrence in artificial neural networks.
- How research in RNNs has lead to state-of-the-art performance on a range of challenging problems.
This was a big post.
Do you have any questions about RNNs for deep learning?
Ask your questions in the comments below, and I will do my best to answer them.
Can you recommend the hardware requirements for the recurrent and lstm networks?
Do you think lstm will perform satisfactorily on a time series with approximately 3000 data points?
Last but not the least, please write a blog post giving the sequence in which one should read your lstm blogs. There are many and it’s hard to decide where to start.
Modest modern hardware. GPUs don’t help much unless you have very large data. In that case I’d recommend renting on AWS.
Do you consider creating a chatbot with LSTMs?
People do, I have not.
can the LSTM model predict random sequence which follow a poisson distribution ?
Generally, neural networks cannot predict a pseudo random series.
How far can LSTMs and RNN go in providing a fool proof algorithm for fake news detection?
Nothing is foolproof.
I think with enough data, that a good model could be developed.
I’m wondering if you had any experience (or knew of any work done) in implementing LSTMs for classifying byte sequences in mainframe (IBM z/OS) data set records with no delimiters within records (analogous to rows in a data table in which you don’t know where columns start and end, but you DO know where rows start and end). If neither, any quick ideas / recommendations would be appreciated as well!
Thanks for this great site!
This is a very specific case and I have not seen it.
It sounds straightforward though. You must convert the inputs into a sequence of integers for the LSTM to be able to learn the problem.
It is similar to mapping words to integers as we do in NLP problems.
Does that help?
here my question is can we say that LSTM and RNN are type of deep learning architecture if so why in some blogs and papers i have read that deep RNN what is the difference? am so confused
LSTM is a type of RNN.
An LSTM or RNN is deep if it has many layers.
In nlp based applications, which of the following network is preferred?
It really depends on the specific NLP task.
I am a phd student. My colleagues and doctors state that: “lstm performs better than Gru in case of nlp application especially in text similarity.”
Can you tell me your opinion about such statement?
I would say it depends on the specific dataset.
Can you tell me when can i use lstm and when can I use gru?
What is the benefit of lstm and gru?
I my opinion, I think the only advantage of the gru is it is faster than the lstm?
You can use LSTM or GRU, typically not both in the same model.
GRU is simpler and faster.
Can Gru gives higher performance and accuracy than lstm?
Depends on the model and the specific dataset.
How can i change the hyperparameter lstm memory cell in keras?
You can only change the number of units or nodes in the LSTM layer.
I really admire you for the efforts you put in to give us such contents.
But after reading this post, I have few doubts. I just want brief answer/points.
What are the types of RNN-
some say its one to one, one to many and more. (reference-https://iq.opengenus.org/types-of-rnn/)
some say its LSTM, GRU and more(reference – https://machinelearningmastery.com/recurrent-neural-network-algorithms-for-deep-learning/)
What are the RNN architectures(if architecture is different from type)-
is architecture diff from types?
What are the types of LSTM-
some say its vanilla, stacked and more( ref-https://machinelearningmastery.com/how-to-develop-lstm-models-for-time-series-forecasting/)
some say its lstm classic, peephole and more(ref-https://blog.exxactcorp.com/5-types-lstm-recurrent-neural-network/)
What are types of LSTM architecture(if architecture is different from type)
i just need names so that i can search them individually and learn.
thanks and keep up the good work.
LSTM and GRU are types of RNN.
There are many architectures, here are some:
Thanks for this great post on RNNs! I want to cite the information you provided here in my research paper. Could you introduce me to some papers on RNN for referencing?
See this re citing:
Thanks for the great tutorial Jason.
Your posts always help me to understand ML/DL deeply.
I have some questions here please.
Can you help me for interpretation of output equation of GRU ? Like what is y_t ?
How can GRU define the size of output (y_t) when implementing ?
Please help me to define and refer some good papers or tutorials in this context if possible.
Sorry, I don’t have tutorials on GRU, but I hope to cover it in the future.