The encoder-decoder model provides a pattern for using recurrent neural networks to address challenging sequence-to-sequence prediction problems such as machine translation. Attention is an extension to the encoder-decoder model that improves the performance of the approach on longer sequences. Global attention is a simplification of attention that may be easier to implement in declarative deep […]
Search results for "transfer learning"
How Does Attention Work in Encoder-Decoder Recurrent Neural Networks
Attention is a mechanism that was developed to improve the performance of the Encoder-Decoder RNN on machine translation. In this tutorial, you will discover the attention mechanism for the Encoder-Decoder model. After completing this tutorial, you will know: About the Encoder-Decoder model and attention mechanism for machine translation. How to implement the attention mechanism step-by-step. […]
How to Scale Data for Long Short-Term Memory Networks in Python
The data for your sequence prediction problem probably needs to be scaled when training a neural network, such as a Long Short-Term Memory recurrent neural network. When a network is fit on unscaled data that has a range of values (e.g. quantities in the 10s to 100s) it is possible for large inputs to slow […]
How to Code a Neural Network with Backpropagation In Python (from scratch)
The backpropagation algorithm is used in the classical feed-forward artificial neural network. It is the technique still used to train large deep learning networks. In this tutorial, you will discover how to implement the backpropagation algorithm for a neural network from scratch with Python. After completing this tutorial, you will know: How to forward-propagate an […]
How To Implement The Perceptron Algorithm From Scratch In Python
The Perceptron algorithm is the simplest type of artificial neural network. It is a model of a single neuron that can be used for two-class classification problems and provides the foundation for later developing much larger networks. In this tutorial, you will discover how to implement the Perceptron algorithm from scratch with Python. After completing […]
How to Train XGBoost Models in the Cloud with Amazon Web Services
The XGBoost library provides an implementation of gradient boosting designed for speed and performance. It is implemented to make best use of your computing resources, including all CPU cores and memory. In this post you will discover how you can setup a server on Amazon’s cloud service to quickly and cheaply create very large models. After […]
Crash Course on Multi-Layer Perceptron Neural Networks
Artificial neural networks are a fascinating area of study, although they can be intimidating when just getting started. There is a lot of specialized terminology used when describing the data structures and algorithms used in the field. In this post, you will get a crash course in the terminology and processes used in the field of multi-layer […]
Non-Linear Classification in R
In this post you will discover 8 recipes for non-linear classification in R. Each recipe is ready for you to copy and paste and modify for your own problem. All recipes in this post use the iris flowers dataset provided with R in the datasets package. The dataset describes the measurements if iris flowers and requires classification of […]
Non-Linear Regression in R
In this post you will discover 4 recipes for non-linear regression in R. There are many advanced methods you can use for non-linear regression, and these recipes are but a sample of the methods you could use. Let’s get started. Each example in this post uses the longley dataset provided in the datasets package that comes with R. The […]