A Gentle Introduction to Positional Encoding in Transformer Models, Part 1

In languages, the order of the words and their position in a sentence really matters. The meaning of the entire sentence can change if the words are re-ordered. When implementing NLP solutions, recurrent neural networks have an inbuilt mechanism that deals with the order of sequences. The transformer model, however, does not use recurrence or convolution and treats each data point as independent of the other. Hence, positional information is added to the model explicitly to retain the information regarding the order of words in a sentence. Positional encoding is the scheme through which the knowledge of the order of objects in a sequence is maintained.

For this tutorial, we’ll simplify the notations used in this remarkable paper, Attention Is All You Need by Vaswani et al. After completing this tutorial, you will know:

  • What is positional encoding, and why it’s important
  • Positional encoding in transformers
  • Code and visualize a positional encoding matrix in Python using NumPy

Kick-start your project with my book Building Transformer Models with Attention. It provides self-study tutorials with working code to guide you into building a fully-working transformer model that can
translate sentences from one language to another...

Let’s get started.

A gentle introduction to positional encoding in transformer models
Photo by Muhammad Murtaza Ghani on Unsplash, some rights reserved

Tutorial Overview

This tutorial is divided into four parts; they are:

  1. What is positional encoding
  2. Mathematics behind positional encoding in transformers
  3. Implementing the positional encoding matrix using NumPy
  4. Understanding and visualizing the positional encoding matrix

What Is Positional Encoding?

Positional encoding describes the location or position of an entity in a sequence so that each position is assigned a unique representation. There are many reasons why a single number, such as the index value, is not used to represent an item’s position in transformer models. For long sequences, the indices can grow large in magnitude. If you normalize the index value to lie between 0 and 1, it can create problems for variable length sequences as they would be normalized differently.

Transformers use a smart positional encoding scheme, where each position/index is mapped to a vector. Hence, the output of the positional encoding layer is a matrix, where each row of the matrix represents an encoded object of the sequence summed with its positional information. An example of the matrix that encodes only the positional information is shown in the figure below.

A Quick Run-Through of the Trigonometric Sine Function

This is a quick recap of sine functions; you can work equivalently with cosine functions. The function’s range is [-1,+1]. The frequency of this waveform is the number of cycles completed in one second. The wavelength is the distance over which the waveform repeats itself. The wavelength and frequency for different waveforms are shown below:


Want to Get Started With Building Transformer Models with Attention?

Take my free 12-day email crash course now (with sample code).

Click to sign-up and also get a free PDF Ebook version of the course.

Positional Encoding Layer in Transformers

Let’s dive straight into this. Suppose you have an input sequence of length $L$ and require the position of the $k^{th}$ object within this sequence. The positional encoding is given by sine and cosine functions of varying frequencies:

\begin{eqnarray}
P(k, 2i) &=& \sin\Big(\frac{k}{n^{2i/d}}\Big)\\
P(k, 2i+1) &=& \cos\Big(\frac{k}{n^{2i/d}}\Big)
\end{eqnarray}

Here:

$k$: Position of an object in the input sequence, $0 \leq k < L/2$

$d$: Dimension of the output embedding space

$P(k, j)$: Position function for mapping a position $k$ in the input sequence to index $(k,j)$ of the positional matrix

$n$: User-defined scalar, set to 10,000 by the authors of Attention Is All You Need.

$i$: Used for mapping to column indices $0 \leq i < d/2$, with a single value of $i$ maps to both sine and cosine functions

In the above expression, you can see that even positions correspond to a sine function and odd positions correspond to cosine functions.

Example

To understand the above expression, let’s take an example of the phrase “I am a robot,” with n=100 and d=4. The following table shows the positional encoding matrix for this phrase. In fact, the positional encoding matrix would be the same for any four-letter phrase with n=100 and d=4.

Coding the Positional Encoding Matrix from Scratch

Here is a short Python code to implement positional encoding using NumPy. The code is simplified to make the understanding of positional encoding easier.

Understanding the Positional Encoding Matrix

To understand the positional encoding, let’s start by looking at the sine wave for different positions with n=10,000 and d=512.

The following figure is the output of the above code:

Sine wave for different position indices

You can see that each position $k$ corresponds to a different sinusoid, which encodes a single position into a vector. If you look closely at the positional encoding function, you can see that the wavelength for a fixed $i$ is given by:

$$
\lambda_{i} = 2 \pi n^{2i/d}
$$

Hence, the wavelengths of the sinusoids form a geometric progression and vary from $2\pi$ to $2\pi n$. The scheme for positional encoding has a number of advantages.

  1. The sine and cosine functions have values in [-1, 1], which keeps the values of the positional encoding matrix in a normalized range.
  2. As the sinusoid for each position is different, you have a unique way of encoding each position.
  3. You have a way of measuring or quantifying the similarity between different positions, hence enabling you to encode the relative positions of words.

Visualizing the Positional Matrix

Let’s visualize the positional matrix on bigger values. Use Python’s matshow() method from the matplotlib library. Setting n=10,000 as done in the original paper, you get the following:

The positional encoding matrix for n=10,000, d=512, sequence length=100

What Is the Final Output of the Positional Encoding Layer?

The positional encoding layer sums the positional vector with the word encoding and outputs this matrix for the subsequent layers. The entire process is shown below.

The positional encoding layer in the transformer

Further Reading

This section provides more resources on the topic if you are looking to go deeper.

Books

Papers

Articles

Summary

In this tutorial, you discovered positional encoding in transformers.

Specifically, you learned:

  • What is positional encoding, and why it is needed.
  • How to implement positional encoding in Python using NumPy
  • How to visualize the positional encoding matrix

Do you have any questions about positional encoding discussed in this post? Ask your questions in the comments below, and I will do my best to answer.

Learn Transformers and Attention!

Building Transformer Models with Attention

Teach your deep learning model to read a sentence

...using transformer models with attention

Discover how in my new Ebook:
Building Transformer Models with Attention

It provides self-study tutorials with working code to guide you into building a fully-working transformer models that can
translate sentences from one language to another...

Give magical power of understanding human language for
Your Projects


See What's Inside

, ,

45 Responses to A Gentle Introduction to Positional Encoding in Transformer Models, Part 1

  1. yuanmu April 13, 2022 at 11:42 pm #

    Thanks for the great explanation!
    Should the range of k be [0, L) instead of [0, L/2)?
    Since the code: for k in range(seq_len)

      • seth G June 22, 2022 at 11:05 am #

        Sorry but I skimmed through the link and I still don’t see why k<L/2.

    • Lucas Thimoteo October 10, 2022 at 11:23 am #

      Hello, I read both links and I think k < L/2 is a mistake. It should be k < L, since k is the index corresponding to the token in the sequence.

      On the other hand, i < d/2 makes total sense because the progression of the dimension is built on 2i and 2i+1.

      • Anonymous January 30, 2023 at 8:47 pm #

        Completely agree with you!

      • Andrei Serebro April 3, 2023 at 2:43 am #

        Yes, that’s right. James, seems you do not try to understand what people in comments are saying to you, I had a feeling that there is no good understanding from your side, or it was just lack of time you were ready to invest into the material you prepared.

  2. Yoan B. June 4, 2022 at 4:26 am #

    Thanks Jason great tutorials !

    I think there’re errors in the trigonometric table section.
    The graphs for sin(2 * 2Pi) and sin(t) go beyond the range [-1:1], either the graph is wrong or the formulas on the left are not the corresponding one.

    • James Carmichael June 4, 2022 at 10:15 am #

      Thank you for the feedback Yoan B!

  3. Tom O. June 4, 2022 at 1:56 pm #

    Very good! Note that I would add plt.show() to avoid head scratching when pasting the examples into ipython.

    • James Carmichael June 5, 2022 at 10:20 am #

      Great feedback Tom!

  4. Shrikant Malviya August 15, 2022 at 11:21 pm #

    “In the above expression we can see that even positions correspond to sine function and odd positions correspond to even positions.”

    Something is wrong or missing in the above statement.

    • James Carmichael August 16, 2022 at 9:47 am #

      Hi Shrikant…Thank you for the feedback! We will review statement in question.

  5. Noman Saleem August 26, 2022 at 10:15 pm #

    Very Nicely Explained. Thanks 🙂

    • James Carmichael August 27, 2022 at 6:08 am #

      You are very welcome! Thank you for your feedback and support Noman!

  6. abraham September 24, 2022 at 12:26 am #

    Hi,
    Is it plausible to use positional encoding for time series prediction with LSTM and Conv1D?

  7. Lucas Thimoteo October 10, 2022 at 11:23 am #

    Hello, I read both links and I think k < L/2 is a mistake. It should be k < L, since k is the index corresponding to the token in the sequence.

    On the other hand, i < d/2 makes total sense because the progression of the dimension is built on 2i and 2i+1.

    • James Carmichael October 11, 2022 at 6:57 am #

      Hi Luca…Thank you for your support and feedback! We will review the content.

  8. Mayank November 6, 2022 at 7:26 am #

    Hi, I have two questions

    1. Do positional embeddings learn just like word embeddings or the embedding values are assigned just based on sine and the cosine graph?

    2. Are positional embedding and word embedding values independent of each other?

  9. A March 10, 2023 at 1:44 am #

    The code doesnt work for an odd embedding vector dimension, the last position would always be left without any assign, could easily be solved with a if statement, but I wonder if odd dimensions for embeddings are even used.

    • James Carmichael March 10, 2023 at 8:05 am #

      Thank you for your feedback A!

  10. Abdi April 7, 2023 at 10:17 pm #

    Excuse me if, for time series forecasting with transformer encoder positional encoding and input, masking is necessary. Because I think in chronicle arranged time series, inputs are ordered in time, and there is no any displacement, also when we use walk forward valodation.

  11. sergiu June 29, 2023 at 2:32 am #

    The positional encoding for example “I am a robot” looks strange for me, same as the output of “getPositionEncoding” function and many questions arise, which makes me more confused. First question, Values are not unique : P01 and P03 or P03 and P13, which one should contain unique values: rows or cols? As I understood the row size = embedding size, and if a row represents the positional encoding for one token, then we can use same values for row. Second question, values are not linear increasing: P00, P10, P20, P30. There is no order, how do we know which one is first, second, … last one if values are not increasing linearly?

    • James Carmichael June 29, 2023 at 8:52 am #

      Hi sergiu…Please rephrase and/or simplify your query if possible so that we may better assist you.

  12. Ryan July 5, 2023 at 10:30 am #

    I think this article is not as good as the others on this website. The concept is not clearly explained. Suggest adding more details and having a good logic between the contents.

    • James Carmichael July 6, 2023 at 8:33 am #

      Thank you Ryan for your feedback and suggestions!

  13. omid October 26, 2023 at 6:30 am #

    Hello,

    Thank you for the explanation.
    I’m having trouble grasping how this method encodes the relative positions of words.
    Is there a formal explanation for this? I understand that for every position “p_i,” it holds that “p_i = T(k) p_i+k,” but I can’t quite comprehend how this concept is beneficial.
    I appreciate your assistance in advance.

  14. Daniel October 30, 2023 at 10:25 pm #

    Do the “Positional vector” and “Positional encoding” functions really take the input words as parameters the way the image under “What Is the Final Output” section implies?

    If not, perhaps it would be a good idea to update the image to make them take input index or something as input instead, because it’s confusing the way it is now. It sort of looks like any position containing the word “am” will get the same positional encoding, and that’s not true, is it?

  15. Giovanni November 24, 2023 at 1:57 pm #

    Io volevo l’amico Jason dai denti a sciabola

  16. Ali Haidar Ahmad December 26, 2023 at 12:10 pm #

    You are the best. I was confused by the explanations that I found on other sites about the math intuition behind the sinusoidal, but now it’s clear.

    • James Carmichael December 27, 2023 at 11:31 am #

      That is great to know Ali! We appreciate your support and feedback!

  17. Anonymous January 4, 2024 at 10:04 pm #

    In the transformer paper and in your tutorial the position encoding is just added (not appended) to the embedding vector, so its dimension stays the same. Doesn’t this “spoil” the working of the attention mechanism, as the tokens are thus modified? I assume it still works well enough, because of the high dimensionality of these vectors. But if you add two high-dimensional vectors, will the sum vector still be sufficiently similar to the embedding vector to not confuse it with others? It will be equally similar to the positional encoding vector. So shouldn’t the latter be smaller, as words are still more important than positions (I guess)? Do you know of any paper discussing this?

  18. afsfg January 29, 2024 at 1:17 am #

    the final step is a sum or a concatenation?

    • James Carmichael January 29, 2024 at 7:01 am #

      Hi afsfg…Please clarify the code sample you are referring to so that we may better assist you.

      • afsfg February 2, 2024 at 6:20 am #

        Thank you. My question was regarding this paragraph:

        What Is the Final Output of the Positional Encoding Layer?
        The positional encoding layer sums the positional vector with the word encoding and outputs this matrix for the subsequent layers. The entire process is shown below.

        I’m curious if “sums” means actually adding two vectors, or it means concatenating two vectors. And why it is adding, not concatenating? what’s the justification of adding two embeddings from two different domains?

        • James Carmichael February 2, 2024 at 10:27 am #

          Hi afsfg…You are very welcome!

          The phrase “sums the positional vector with the word encoding and outputs this matrix for the subsequent layers” refers to a process commonly used in the architecture of transformer models, like those underlying many modern natural language processing (NLP) systems.

          Let’s break it down:

          1. **Word Encoding (Word Embeddings):** Each word (or token) in a sentence is converted into a numerical form, known as an embedding. This embedding captures the semantic meaning of the word in a high-dimensional space. Essentially, similar words have similar embeddings. These embeddings are learned from data and are an integral part of neural network models for language tasks.

          2. **Positional Encoding:** Since the transformer architecture does not inherently process sequential data in order (unlike RNNs or LSTMs), it requires a method to understand the order of words in a sentence. Positional encoding is added to solve this problem. It involves generating another set of vectors that encode the position of each word in the sentence. Each positional vector is unique to its position, ensuring that the model can recognize the order of words. The positional encoding is designed so that it can be combined with the word embeddings, usually through addition, without losing the information contained in either.

          3. **Summing the Positional Vector with the Word Encoding:** The process involves element-wise addition of the positional encoding vector to the word embedding vector for each word. This combined vector now contains both the semantic information of the word (from the word embedding) and its position in the sentence (from the positional encoding). This is crucial for the model to understand both the meaning of words and how the order of words affects the meaning of the sentence.

          4. **Outputting this Matrix for the Subsequent Layers:** The resulting matrix, after summing positional vectors and word embeddings, is then passed through the subsequent layers of the transformer model. These layers include self-attention mechanisms and fully connected networks, which process this combined information to perform tasks such as language understanding, translation, or generation. The matrix effectively serves as the input for these complex operations, allowing the model to consider both the meaning of individual words and their context within the sentence.

          This process is fundamental to the functioning of transformer models, enabling them to achieve state-of-the-art performance on a wide range of NLP tasks.

  19. YG March 27, 2024 at 5:17 pm #

    I think the sentence
    “you can see that the wavelength for a fixed i is given by: …”

    should be
    “you can see that the wavelength for a fixed ‘T’ is given by: …”

    and as many already pointed out, K range should be below:
    0 <= K < L

  20. Pavan Kumar April 20, 2024 at 3:23 pm #

    For example, let’s consider a sequence of 5 tokens, where each token is represented by an embedding of 10 dimensions; thus, each token’s embedding shape is 1×10. Instead of directly adding the raw positions 0 to 4 to the embeddings, which might overly alter the original semantic embeddings, a sinusoidal function is utilized. This function takes as inputs the token’s position in the sequence, the index within the embedding dimension, and the total number of dimensions, producing a 1×10 positional embedding for each token.

    These positional embeddings are then added to the original token embeddings. In the Transformer architecture, during the calculation of Query, Key, and Value vectors, the matrix multiplication in the attention mechanism tends to assign higher scores to adjacent tokens, unless there is a strong semantic signal from the embeddings that suggests otherwise. This maintains the influence of token proximity unless overridden by significant contextual differences embedded in the tokens.

    • James Carmichael April 21, 2024 at 10:27 am #

      Thank you for your feedback Pavan! You’ve nicely outlined the concept of positional embeddings in the Transformer model architecture and how they influence the self-attention mechanism. Let’s delve a bit deeper into each of these components and their operational dynamics within the Transformer framework:

      ### Positional Embeddings

      1. **Purpose**: In Transformer models, positional embeddings provide necessary information about the order of tokens in the input sequence. Since the self-attention mechanism by itself doesn’t inherently process sequential data in order (i.e., it treats input as a set of tokens independently of their positions), positional embeddings are crucial for incorporating the sequence order.

      2. **Sinusoidal Function**: The positional encoding formula uses sine and cosine functions of different frequencies:

      \[
      PE_{(pos, 2i)} = \sin\left(\frac{pos}{10000^{2i/d_{\text{model}}}}\right)
      \]
      \[
      PE_{(pos, 2i+1)} = \cos\left(\frac{pos}{10000^{2i/d_{\text{model}}}}\right)
      \]

      Where:
      – \( pos \) is the position of the token in the sequence.
      – \( i \) is the dimension index.
      – \( d_{\text{model}} \) is the total number of dimensions in the embedding.
      – This generates a unique positional embedding for each position, with the pattern allowing the model to learn to attend based on relative positions, as the distance between any two positions can be encoded into the learned weights.

      3. **Integration with Token Embeddings**: These positional embeddings are element-wise added to the token embeddings. This way, each token’s representation reflects not only its own inherent meaning but also its position in the sequence.

      ### Self-Attention Mechanism

      1. **Query, Key, Value Vectors**: In the Transformer block, embeddings (integrated with positional information) are transformed into three vectors: Queries, Keys, and Values, which are used in the attention mechanism.

      2. **Attention Calculation**: The core of the attention mechanism can be described by the formula:

      \[
      \text{Attention}(Q, K, V) = \text{softmax}\left(\frac{QK^T}{\sqrt{d_k}}\right)V
      \]

      – \( Q, K, V \) are the matrices of queries, keys, and values.
      – \( d_k \) is the dimension of the keys.
      – The softmax score determines the degree to which each value is factored into the output based on the dot-product similarity between corresponding query and key.

      3. **Role of Proximity and Contextual Similarity**: While the positional embeddings encourage the model to consider adjacent tokens more heavily (since their positional encodings are more similar), the ultimate attention each token receives is determined by a combination of its semantic similarity (as reflected in the query-key relationships) and its relative position. This dual consideration allows the Transformer to maintain context sensitivity, enhancing its ability to handle a wide variety of language understanding tasks.

      ### Summary

      By combining positional embeddings with the token embeddings, Transformers maintain awareness of both the individual token’s meaning and its position in the sequence. This allows for effective modeling of language where the meaning often depends significantly on word order and contextual proximity. The sinusoidal pattern of positional encodings also ensures that the model can generalize to sequences of different lengths and recognize patterns across various positions within the data.

Leave a Reply

Machine Learning Mastery is part of Guiding Tech Media, a leading digital media publisher focused on helping people figure out technology. Visit our corporate website to learn more about our mission and team.