Sentiment analysis is a natural language processing problem where text is understood, and the underlying intent is predicted. In this post, you will discover how you can predict the sentiment of movie reviews as either positive or negative in Python using the Keras deep learning library. After reading this post, you will know: About the […]
Search results for "word embedding"
Start Here with Machine Learning
Need Help Getting Started with Applied Machine Learning? These are the Step-by-Step Guides that You’ve Been Looking For! What do you want help with? The most common question I’m asked is: “how do I get started?” My best advice for getting started in machine learning is broken down into a 5-step process: Step 1: Adjust Mindset. […]
Generate Realistic and Consistent Faces in Stable Diffusion
The deep learning model of Stable Diffusion is huge. The weight file is multiple GB large. Retraining the model means to update a lot of weights and that is a lot of work. Sometimes we must modify the Stable Diffusion model, for example, to define a new interpretation of prompts or make the model to […]
Prompting Techniques for Stable Diffusion
Generating pictures using Stable Diffusion in all cases would involve to submit a prompt to the pipeline. This is only one of the parameters, but the most important one. An incomplete or poorly constructed prompt would make the resulting image not as you would expect. In this post, you will learn some key techniques to […]
How to Create Images Using Stable Diffusion Web UI
Launching the Stable Diffusion Web UI can be done in one command. After that, you can control the image generation pipeline from a browser. The pipeline has a lot of moving parts and all are important in one way or another. To effectively command Stable Diffusion to generate images, you should recognize the widgets from […]
Brief Introduction to Diffusion Models for Image Generation
The advance of generative machine learning models makes computers capable of creative work. In the scope of drawing pictures, there are a few notable models that allow you to convert a textual description into an array of pixels. The most powerful models today are part of the family of diffusion models. In this post, you […]
What are Large Language Models
Large language models (LLMs) are recent advances in deep learning models to work on human languages. Some great use case of LLMs has been demonstrated. A large language model is a trained deep-learning model that understands and generates text in a human-like fashion. Behind the scene, it is a large transformer model that does all […]
Training the Transformer Model
We have put together the complete Transformer model, and now we are ready to train it for neural machine translation. We shall use a training dataset for this purpose, which contains short English and German sentence pairs. We will also revisit the role of masking in computing the accuracy and loss metrics during the training […]
A Gentle Introduction to Positional Encoding in Transformer Models, Part 1
In languages, the order of the words and their position in a sentence really matters. The meaning of the entire sentence can change if the words are re-ordered. When implementing NLP solutions, recurrent neural networks have an inbuilt mechanism that deals with the order of sequences. The transformer model, however, does not use recurrence or […]
The Transformer Model
We have already familiarized ourselves with the concept of self-attention as implemented by the Transformer attention mechanism for neural machine translation. We will now be shifting our focus to the details of the Transformer architecture itself to discover how self-attention can be implemented without relying on the use of recurrence and convolutions. In this tutorial, […]