# Basics of Mathematical Notation for Machine Learning

You cannot avoid mathematical notation when reading the descriptions of machine learning methods.

Often, all it takes is one term or one fragment of notation in an equation to completely derail your understanding of the entire procedure. This can be extremely frustrating, especially for machine learning beginners coming from the world of development.

You can make great progress if you know a few basic areas of mathematical notation and some tricks for working through the description of machine learning methods in papers and books.

In this tutorial, you will discover the basics of mathematical notation that you may come across when reading descriptions of techniques in machine learning.

After completing this tutorial, you will know:

• Notation for arithmetic, including variations of multiplication, exponents, roots, and logarithms.
• Notation for sequences and sets including indexing, summation, and set membership.
• 5 Techniques you can use to get help if you are struggling with mathematical notation.

Kick-start your project with my new book Linear Algebra for Machine Learning, including step-by-step tutorials and the Python source code files for all examples.

Let’s get started.

• Update May/2018: Added images for some notations to make the explanations clearer. Basics of Mathematical Notation for Machine Learning
Photo by Christian Collins, some rights reserved.

## Tutorial Overview

This tutorial is divided into 7 parts; they are:

1. The Frustration with Math Notation
2. Arithmetic Notation
3. Greek Alphabet
4. Sequence Notation
5. Set Notation
6. Other Notation
7. Getting More Help

Are there other areas of basic math notation required for machine learning that you think I missed?
Let me know in the comments below.

### Need help with Linear Algebra for Machine Learning?

Take my free 7-day email crash course now (with sample code).

Click to sign-up and also get a free PDF Ebook version of the course.

## The Frustration with Math Notation

For example, notation may be used to:

• Describe an algorithm.
• Describe data preparation.
• Describe results.
• Describe a test harness.
• Describe implications.

These descriptions may be in research papers, textbooks, blog posts, and elsewhere.

Often the terms are well defined, but there are also mathematical notation norms that you may not be familiar with.

All it takes is one term or one equation that you do not understand and your understanding of the entire method will be lost. I’ve suffered this problem myself many times, and it is incredibly frustrating!

In this tutorial, we will review some basic mathematical notation that will help you when reading descriptions of machine learning methods.

## Arithmetic Notation

In this section, we will go over some less obvious notations for basic arithmetic as well as a few concepts you may have forgotten since school.

### Simple Arithmetic

The notation for basic arithmetic is as you would write it. For example:

• Addition: 1 + 1 = 2
• Subtraction: 2 – 1 = 1
• Multiplication: 2 x 2 = 4
• Division: 2 / 2 = 1

Most mathematical operations have a sister operation that performs the inverse operation; for example, subtraction is the inverse of addition and division is the inverse of multiplication.

### Algebra

We often want to describe operations abstractly to separate them from specific data or specific implementations.

For this reason we see heavy use of algebra: that is uppercase and/or lowercase letters or words to represents terms or concepts in mathematical notation. It is also common to use letters from the Greek alphabet.

Each sub-field of math may have reserved letters: that is terms or letters that always mean the same thing. Nevertheless, algebraic terms should be defined as part of the description and if they are not, it may just be a poor description, not your fault.

### Multiplication Notation

Multiplication is a common notation and has a few short hands.

Often a little “x” or an asterisk “*” is used to represent multiplication:

You may see a dot notation used; for example:

Which is the same as:

Alternately, you may see no operation and no white space separation between previously defined terms; for example:

Which again is the same thing.

### Exponents and Square Roots

An exponent is a number raised to a power.

The notation is written as the original number, or the base, with a second number, or the exponent, shown as a superscript; for example:

Which would be calculated as 2 multiplied by itself 3 times, or cubing:

A number raised to the power 2 to is said to be its square.

The square of a number can be inverted by calculating the square root. This is shown using the notation of a number and with a tick above, I will use the “sqrt()” function here for simplicity. Here, we know the result and the exponent and we wish to find the base.

In fact, the root operation can be used to inverse any exponent, it just so happens that the default square root assumes an exponent of 2, represented by a subscript 2 in front of the square root tick.

For example, we can invert the cubing of a number by taking the cube root (note, the 3 is not a multiplication here, it is notation before the tick of the root sign):

### Logarithms and e

When we raise 10 to an integer exponent, we often call this an order of magnitude.

Another way to reverse this operation is by calculating the logarithm of the result 100 assuming a base of 10; in notation this is written as log10().

Here, we know the result and the base and wish to find the exponent.

This allows us to move up and down orders of magnitude very easily. Taking the logarithm assuming the base of 2 is also commonly used, given the use of binary arithmetic used in computers. For example:

Another popular logarithm is to assume the natural base called e. The e is reserved and is a special number or a constant called Euler’s number (pronounced “oy-ler“) that refers to a value with practically infinite precision.

Raising e to a power is called a natural exponential function:

It can be inverted using the natural logarithm, which is denoted as ln():

Without going into detail, the natural exponent and natural logarithm prove useful throughout mathematics to abstractly describe the continuous growth of some systems, e.g. systems that grow exponentially such as compound interest.

## Greek Alphabet

Greek letters are used throughout mathematical notation for variables, constants, functions, and more.

For example, in statistics we talk about the mean using the lowercase Greek letter mu, and the standard deviation as the lowercase Greek letter sigma. In linear regression, we talk about the coefficients as the lowercase letter beta. And so on.

It is useful to know all of the uppercase and lowercase Greek letters and how to pronounce them.

When I was a grad student, I printed the Greek alphabet and stuck it on my computer monitor so that I could memorize it. A useful trick!

Below is the full Greek alphabet. Greek Alphabet, Taken from Wikipedia

The Wikipedia page titled “Greek letters used in mathematics, science, and engineering” is also a useful guide as it lists common uses for each Greek letter in different sub-fields of math and science.

## Sequence Notation

Machine learning notation often describes an operation on a sequence.

A sequence may be an array of data or a list of terms.

### Indexing

A key to reading notation for sequences is the notation of indexing elements in the sequence.

Often the notation will specify the beginning and end of the sequence, such as 1 to n, where n will be the extent or length of the sequence.

Items in the sequence are index by a variable such as i, j, k as a subscript. This is just like array notation.

For example, a_i is the i^th element of the sequence a.

If the sequence is two dimensional, two indices may be used; for example:

b_{i,j} is the i,j^th element of the sequence b.

### Sequence Operations

Mathematical operations can be performed over a sequence.

Two operations are performed on sequences so often that they have their own shorthand: the sum and the multiplication.

#### Sequence Summation

The sum over a sequence is denoted as the uppercase Greek letter sigma. It is specified with the variable and start of the sequence summation below the sigma (e.g. i = 1) and the index of the end of the summation above the sigma (e.g. n). This is the sum of the sequence a starting at element 1 to element n.

#### Sequence Multiplication

The multiplication over a sequence is denoted as the uppercase Greek letter pi. It is specified in the same way as the sequence summation with the beginning and end of the operation below and above the letter respectively. This is the product of the sequence a starting at element 1 to element n.

## Set Notation

A set is a group of unique items.

We may see set notation used when defining terms in machine learning.

### Set of Numbers

A common set you may see is a set of numbers, such as a term defined as being within the set of integers or the set of real numbers.

Some common sets of numbers you may see include:

• Set of all natural numbers: N
• Set of all integers: Z
• Set of all real numbers: R

There are other sets; see Special sets on Wikipedia.

We often talk about real values or real numbers when defining terms rather than floating point values, which are really discrete creations for operations in computers.

## Set Membership

It is common to see set membership in definitions of terms.

Set membership is denoted as a symbol that looks like an uppercase “E”. Which means a is defined as being a member of the set R or the set of real numbers.

There is also a host of set operations; two common set operations include:

• Union, or aggregation: A U B
• Intersection, or overlap: A ^ B

## Other Notation

There is other notation that you may come across.

I try to lay some of it out in this section.

It is common to define a method in the abstract and then define it again as a specific implementation with separate notation.

For example, if we are estimating a variable x, we may represent it using a notation that modifies the x; for example: The same notation may have a different meaning in a different context, such as use on different objects or sub-fields of mathematics. For example, a common point of confusion is |x|, which, depending on context, can mean:

• |x|: The absolute or positive value of x.
• |x|: The length of the vector x.
• |x|: The cardinality of the set x.

This tutorial only covered the basics of mathematical notation. There are some subfields of mathematics that are more relevant to machine learning and should be reviewed in more detail. They are:

And perhaps a little bit of multivariate analysis and information theory.

Are there areas of mathematical notation that you think are missing from this post?
Let me know in the comments below.

## 5 Tips for Getting Help with Math Notation

This section lists some tips that you can use when you are struggling with mathematical notation in machine learning.

People wrote the paper or book you are reading.

People that can make mistakes, make omissions, and even make things confusing because they don’t fully understand what they are writing.

Relax the constraints of the notation you are reading slightly and think about the intent of the author. What are they trying to get across?

Perhaps you can even contact the author via email, Twitter, Facebook, LinkedIn, etc., and seek clarification. Remember that academics want other people to understand and use their work (mostly).

### Check Wikipedia

Wikipedia has lists of notation which can help narrow down on the meaning or intent of the notation you are reading.

Two places I recommend you start are:

### Sketch in Code

Mathematical operations are just functions on data.

Map everything you’re reading to pseudocode with variables, for-loops, and more.

You might want to use a scripting language as you go, along with small arrays of contrived data or even an Excel spreadsheet.

As your reading and understanding of the technique improves, your code-sketch of the technique will make more sense, and at the end you will have a mini prototype to play with.

I never used to take much stock in this approach until I saw an academic sketch out a very complex paper in a few lines of matlab with some contrived data. It knocked my socks off because I believed the system had to be coded completely and run with a “real” dataset and that the only option was to get the original code and data. I was very wrong. Also, looking back, the guy was gifted.

I now use this method all the time and sketch techniques in Python.

### Seek Alternatives

There is a trick I use when I’m trying to understand a new technique.

I find and read all the papers that reference the paper I’m reading with the new technique.

Reading other academics interpretation and re-explanation of the technique can often clarify my misunderstandings in the original description.

Not always though. Sometimes it can muddy the waters and introduce misleading explanations or new notation. But more often than not, it helps. After circling back to the original paper and re-reading it, I can often find cases where subsequent papers have actually made errors and misinterpretations of the original method.

### Post a Question

There are places online where people love to explain math to others. Seriously!

Consider taking a screenshot of the notation you are struggling with, write out the full reference or link to it, and post it and your area of misunderstanding to a question-and-answer site.

Two great places to start are:

What are your tricks for working through mathematical notation?
Let me know in the comments below?

This section provides more resources on the topic if you are looking to go deeper.

## Summary

In this tutorial, you discovered the basics of mathematical notation that you may come across when reading descriptions of techniques in machine learning.

Specifically, you learned:

• Notation for arithmetic, including variations of multiplication, exponents, roots, and logarithms.
• Notation for sequences and sets, including indexing, summation, and set membership.
• 5 Techniques you can use to get help if you are struggling with mathematical notation.

Are you struggling with mathematical notation?

Did any of the notation or tips in this post help?
Let me know in the comments below.

## Get a Handle on Linear Algebra for Machine Learning! #### Develop a working understand of linear algebra

...by writing lines of code in python

Discover how in my new Ebook:
Linear Algebra for Machine Learning

It provides self-study tutorials on topics like:
Vector Norms, Matrix Multiplication, Tensors, Eigendecomposition, SVD, PCA and much more...

### 28 Responses to Basics of Mathematical Notation for Machine Learning

1. Emaliana Kasmuri March 26, 2018 at 12:26 pm #

This is a great article. I’ve been struggling to understand maths symbols used in equations. It is a relief to know that I am not alone and you have simplify it. Thank you for writing this.

• Jason Brownlee March 26, 2018 at 2:27 pm #

Thanks, I hope it helps.

• mayur barbhaya January 24, 2020 at 10:57 pm #

thanks for giving us what we trying to find since 5 days, which is actually in depth for new comers in fields of machine learning

• Jason Brownlee January 25, 2020 at 8:36 am #

You’re welcome.

2. Timothy Man March 26, 2018 at 7:20 pm #

Nice summary.

However, keep in mind that in the section of Multiplication Notation your discussion on all notations being equal is only true for scalars. Since we are using vectors and matrices a lot in Machine Learning, we need to be careful in distinguishing dot- and cross-products, which uses unsurprisingly the dot- and cross-notations.

• Jason Brownlee March 27, 2018 at 6:34 am #

Thanks Timothy.

3. Wazir Ali May 24, 2018 at 2:27 am #

your article is very helpful in order to understand basic mathematical notations frequently used in machine learning. Thank you Jason Brownlee

• Jason Brownlee May 24, 2018 at 8:18 am #

Thanks.

4. Mohamed May 31, 2018 at 11:04 am #

An amazing article indeed! Thank you so much for writing this. This is extremely helpful.

• Jason Brownlee May 31, 2018 at 2:12 pm #

5. Paul September 30, 2018 at 12:53 am #

Great article, I wish i’d found it before I started my classes!

• Jason Brownlee September 30, 2018 at 6:03 am #

Thanks Paul!

6. Rob May 21, 2019 at 11:39 am #

Thanks Jason for a great overview – certainly has helped me piece together the underlying notation I’m studying

• Jason Brownlee May 21, 2019 at 2:43 pm #

Thanks, I’m happy to hear that.

7. Leo Goes June 11, 2019 at 4:15 am #

Thanks Mr Jason, great article. Don’t know about you, but 3sqrt(x) representing cubic root of x made my stomach hurt, oh god why

• Jason Brownlee June 11, 2019 at 8:03 am #

It sucks I know, the latex makes it clearer.

Do you have a better way to represent it in plain ascii?

• Leo Goes June 11, 2019 at 12:45 pm #

This is true, in latex it is easier to notice the difference. it’s 8731 in ascii (extended) and 221B in unicode for cubic root (just checked, it works)

• Jason Brownlee June 11, 2019 at 2:23 pm #

Nice!

8. Noah Caldwell July 3, 2019 at 2:24 am #

I’m having trouble figuring out the meaning of a notation I can only describe as ‘fancy letters’ – they’re not Greek, nor do they seem to just be arbitrary variables – they’re upper-case letters that just seem to be written in a different font and have some sort of meaning that I can’t figure out.

For context, I’m reading an overview of Google’s GNMT system here – https://arxiv.org/pdf/1609.08144.pdf – on page 8 on the training criteria, and there’re several examples of this (i.e. the D in the first notation, the O in the second, and the Y in the subscript of the double summation in the third). I’m fairly confident that fancy D means dataset, fancy O might mean output, and I have no idea for the Y – and I’m also pretty confused about how they fit into the equations. Any help/guidance/further reading you could give would be appreciated – thanks!

• Jason Brownlee July 3, 2019 at 8:40 am #

Yes, D is defined as the dataset, O is the objective function.

Don’t worry, it is common to spend a week or weeks coming to terms with one paper. I often find reading the code clearer in many cases – if an open source implementation is provided.

Perhaps you can find others also interested in picking apart the paper and work through it together. Check the subreddit or even people asking questions about it on stackexchange network.

9. R January 18, 2020 at 12:57 am #

Thanks for the information. In general, how are the prime, hat, tilde, symbols used in machine learning equations, etc.? y hat is the predicted value or values, but x predicts y so what is x hat? Prime, tilde examples?

• Jason Brownlee January 18, 2020 at 8:49 am #

Tough question!

It really depends.

Frankly, a paper/book can adopt any notation it wants as long is it stays internally consistent.

10. mayur barbhaya January 24, 2020 at 10:58 pm #

again thank you so much for this content sir.

• Jason Brownlee January 25, 2020 at 8:36 am #

You’re welcome.

11. Aaron Speedy October 14, 2020 at 7:23 am #

Great Article. This has really helped me review what all those symbols mean while I am studying machine learning

• Jason Brownlee October 14, 2020 at 7:38 am #

Thanks, I’m happy to hear that!

• Mark December 11, 2022 at 12:46 pm #

Take common mathematical statements found in Machine Learning and interpret them. What you provided is correct but too basic.

mark

• James Carmichael December 12, 2022 at 9:55 am #

Thank you Mark for your feedback!