Search results for "translation"

kmeans_class_cover

K-Means Clustering for Image Classification Using OpenCV

In a previous tutorial, we have explored the use of the k-means clustering algorithm as an unsupervised machine learning technique that seeks to group similar data into distinct clusters, to uncover patterns in the data.  We have, so far, seen how to apply the k-means clustering algorithm to a simple two-dimensional dataset containing distinct clusters, […]

Continue Reading
Maximizing Productivity with ChatGPT

Maximizing Productivity with ChatGPT

Maximizing Productivity with ChatGPT Let Generative AI Help You Work Smarter Why Are Large Language Models So Powerful? …the secret is “Pattern Recognition and Understanding“ Large Language Models (LLMs), like ChatGPT, are incredibly powerful because they learn to understand and generate human language patterns. This capability goes beyond merely memorizing phrases; instead, they learn the […]

Continue Reading
00003-4136881981-In a library

Get a Taste of LLMs from GPT4All

Large language models have become popular recently. ChatGPT is fashionable. Trying out ChatGPT to understand what LLMs are about is easy, but sometimes, you may want an offline alternative that can run on your computer. In this post, you will learn about GPT4All as an LLM that you can install on your computer. In particular, […]

Continue Reading
00002-2320356430-A table outdo

What are Large Language Models

Large language models (LLMs) are recent advances in deep learning models to work on human languages. Some great use case of LLMs has been demonstrated. A large language model is a trained deep-learning model that understands and generates text in a human-like fashion. Behind the scene, it is a large transformer model that does all […]

Continue Reading
norbert-braun-uU8n5LuzpTc-unsplash

Building Transformer Models with Attention Crash Course. Build a Neural Machine Translator in 12 Days

Transformer is a recent breakthrough in neural machine translation. Natural languages are complicated. A word in one language can be translated into multiple words in another, depending on the context. But what exactly a context is, and how you can teach the computer to understand the context was a big problem to solve. The invention […]

Continue Reading
samet-erkoseoglu-B0nUaoWnr0M-unsplash

A Brief Introduction to BERT

As we learned what a Transformer is and how we might train the Transformer model, we notice that it is a great tool to make a computer understand human language. However, the Transformer was originally designed as a model to translate one language to another. If we repurpose it for a different task, we would […]

Continue Reading
karsten-wurth-algc0FKHeMA-unsplash

Inferencing the Transformer Model

We have seen how to train the Transformer model on a dataset of English and German sentence pairs and how to plot the training and validation loss curves to diagnose the model’s learning performance and decide at which epoch to run inference on the trained model. We are now ready to run inference on the […]

Continue Reading
training_cover

Training the Transformer Model

We have put together the complete Transformer model, and now we are ready to train it for neural machine translation. We shall use a training dataset for this purpose, which contains short English and German sentence pairs. We will also revisit the role of masking in computing the accuracy and loss metrics during the training […]

Continue Reading