Practical Recommendations for Deep Learning Neural Network Practitioners

Recommendations for Deep Learning Neural Network Practitioners

Deep learning neural networks are relatively straightforward to define and train given the wide adoption of open source libraries. Nevertheless, neural networks remain challenging to configure and train. In his 2012 paper titled “Practical Recommendations for Gradient-Based Training of Deep Architectures” published as a preprint and a chapter of the popular 2012 book “Neural Networks: […]

Continue Reading
8 Tricks for Configuring Backpropagation to Train Better Neural Networks, Faster

8 Tricks for Configuring Backpropagation to Train Better Neural Networks

Neural network models are trained using stochastic gradient descent and model weights are updated using the backpropagation algorithm. The optimization solved by training a neural network model is very challenging and although these algorithms are widely used because they perform so well in practice, there are no guarantees that they will converge to a good […]

Continue Reading
A Gentle Introduction to the Challenge of Training Deep Learning Neural Network Models

A Gentle Introduction to the Challenge of Training Deep Learning Neural Network Models

Deep learning neural networks learn a mapping function from inputs to outputs. This is achieved by updating the weights of the network in response to the errors the model makes on the training dataset. Updates are made to continually reduce this error until either a good enough model is found or the learning process gets […]

Continue Reading
Loss and Accuracy Learning Curves on the Train and Test Sets for an MLP on Problem 1

How to Improve Performance With Transfer Learning for Deep Learning Neural Networks

An interesting benefit of deep learning neural networks is that they can be reused on related problems. Transfer learning refers to a technique for predictive modeling on a different but somehow similar problem that can then be reused partly or wholly to accelerate the training and improve the performance of a model on the problem […]

Continue Reading
How to Avoid Exploding Gradients in Neural Networks With Gradient Clipping

How to Avoid Exploding Gradients With Gradient Clipping

Training a neural network can become unstable given the choice of error function, learning rate, or even the scale of the target variable. Large updates to weights during training can cause a numerical overflow or underflow often referred to as “exploding gradients.” The problem of exploding gradients is more common with recurrent neural networks, such […]

Continue Reading