The Keras library provides a way to calculate and report on a suite of standard metrics when training deep learning models. In addition to offering standard metrics for classification and regression problems, Keras also allows you to define and report on your own custom metrics when training deep learning models. This is particularly useful if […]

# Archive | Deep Learning

Deep Learning

## 10 Command Line Recipes for Deep Learning on Amazon Web Services

Running large deep learning processes on Amazon Web Services EC2 is a cheap and effective way to learn and develop models. For just a few dollars you can get access to tens of gigabytes of RAM, tens of CPU cores, and multiple GPUs. I highly recommend it. If you are new to EC2 or the […]

## 9 Ways to Get Help with Deep Learning in Keras

Keras is a Python deep learning library that can use the efficient Theano or TensorFlow symbolic math libraries as a backend. Keras is so easy to use that you can develop your first Multilayer Perceptron, Convolutional Neural Network, or LSTM Recurrent Neural Network in minutes. You may have technical questions when you get started using […]

## A Gentle Introduction to Mini-Batch Gradient Descent and How to Configure Batch Size

Stochastic gradient descent is the dominant method used to train deep learning models. There are three main variants of gradient descent and it can be confusing which one to use. In this post, you will discover the one type of gradient descent you should use in general and how to configure it. After completing this […]

## Gentle Introduction to the Adam Optimization Algorithm for Deep Learning

The choice of optimization algorithm for your deep learning model can mean the difference between good results in minutes, hours, and days. The Adam optimization algorithm is an extension to stochastic gradient descent that has recently seen broader adoption for deep learning applications in computer vision and natural language processing. In this post, you will […]

## How to Get Reproducible Results with Keras

Neural network algorithms are stochastic. This means they make use of randomness, such as initializing to random weights, and in turn the same network trained on the same data can produce different results. This can be confusing to beginners as the algorithm appears unstable, and in fact they are by design. The random initialization allows […]

## How to Evaluate the Skill of Deep Learning Models

I often see practitioners expressing confusion about how to evaluate a deep learning model. This is often obvious from questions like: What random seed should I use? Do I need a random seed? Why don’t I get the same results on subsequent runs? In this post, you will discover the procedure that you can use […]

## Exploratory Configuration of a Multilayer Perceptron Network for Time Series Forecasting

It can be difficult when starting out on a new predictive modeling project with neural networks. There is so much to configure, and no clear idea where to start. It is important to be systematic. You can break bad assumptions and quickly hone in on configurations that work and areas for further investigation likely to […]

## How To Improve Deep Learning Performance

20 Tips, Tricks and Techniques That You Can Use To Fight Overfitting and Get Better Generalization How can you get better performance from your deep learning model? It is one of the most common questions I get asked. It might be asked as: How can I improve accuracy? …or it may be reversed as: What […]

## What You Know About Deep Learning Is A Lie

Getting started in deep learning is a struggle. It’s a struggle because deep learning is taught by academics, for academics. If you’re a developer (or practitioner), you’re different. You want results. The way practitioners learn new technologies is by developing prototypes that deliver value quickly. This is a top-down approach to learning, but it is not the way […]