A What programming language should I use for machine learning? The specific programming language or platform that you use does not matter. I strongly believe that the best thing to focus on is how to work through machine learning problems end-to-end (learn more). That being said, I think if you’re not a strong programmer, that […]
Search results for "language model"

How to Develop a Multichannel CNN Model for Text Classification
A standard deep learning model for text classification and sentiment analysis uses a word embedding layer and one-dimensional convolutional neural network. The model can be expanded by using multiple parallel convolutional neural networks that read the source document using different kernel sizes. This, in effect, creates a multichannel convolutional neural network for text that reads […]

How to Implement a Beam Search Decoder for Natural Language Processing
Natural language processing tasks, such as caption generation and machine translation, involve generating sequences of words. Models developed for these problems often operate by generating probability distributions across the vocabulary of output words and it is up to decoding algorithms to sample the probability distributions to generate the most likely sequences of words. In this […]

How to Configure an Encoder-Decoder Model for Neural Machine Translation
The encoder-decoder architecture for recurrent neural networks is achieving state-of-the-art results on standard machine translation benchmarks and is being used in the heart of industrial translation services. The model is simple, but given the large amount of data required to train it, tuning the myriad of design decisions in the model in order get top […]

Encoder-Decoder Recurrent Neural Network Models for Neural Machine Translation
The encoder-decoder architecture for recurrent neural networks is the standard neural machine translation method that rivals and in some cases outperforms classical statistical machine translation methods. This architecture is very new, having only been pioneered in 2014, although, has been adopted as the core technology inside Google’s translate service. In this post, you will discover […]

Caption Generation with the Inject and Merge Encoder-Decoder Models
Caption generation is a challenging artificial intelligence problem that draws on both computer vision and natural language processing. The encoder-decoder recurrent neural network architecture has been shown to be effective at this problem. The implementation of this architecture can be distilled into inject and merge based models, and both make different assumptions about the role […]

Encoder-Decoder Models for Text Summarization in Keras
Text summarization is a problem in natural language processing of creating a short, accurate, and fluent summary of a source document. The Encoder-Decoder recurrent neural network architecture developed for machine translation has proven effective when applied to the problem of text summarization. It can be difficult to apply this architecture in the Keras deep learning […]

How to Use Small Experiments to Develop a Caption Generation Model in Keras
Caption generation is a challenging artificial intelligence problem where a textual description must be generated for a photograph. It requires both methods from computer vision to understand the content of the image and a language model from the field of natural language processing to turn the understanding of the image into words in the right […]

A Gentle Introduction to Deep Learning Caption Generation Models
Caption generation is the challenging artificial intelligence problem of generating a human-readable textual description given a photograph. It requires both image understanding from the domain of computer vision and a language model from the field of natural language processing. It is important to consider and test multiple ways to frame a given predictive modeling problem […]

Deep Learning For Natural Language Processing
Deep Learning for Natural Language Processing Develop Deep Learning Models for your Natural Language Problems Working with Text is… important, under-discussed, and HARD We are awash with text, from books, papers, blogs, tweets, news, and increasingly text from spoken utterances. Every day, I get questions asking how to develop machine learning models for text data. Working […]