Search results for "Principal Component"

Singular Value Decomposition for Dimensionality Reduction in Python

Singular Value Decomposition for Dimensionality Reduction in Python

Reducing the number of input variables for a predictive model is referred to as dimensionality reduction. Fewer input variables can result in a simpler predictive model that may have better performance when making predictions on new data. Perhaps the more popular technique for dimensionality reduction in machine learning is Singular Value Decomposition, or SVD for […]

Continue Reading
A Gentle Introduction to Dimensionality Reduction for Machine Learning

Introduction to Dimensionality Reduction for Machine Learning

The number of input variables or features for a dataset is referred to as its dimensionality. Dimensionality reduction refers to techniques that reduce the number of input variables in a dataset. More input features often make a predictive modeling task more challenging to model, more generally referred to as the curse of dimensionality. High-dimensionality statistics […]

Continue Reading
How to Predict the Probability of Fraudulent Credit Card Transactions

Imbalanced Classification with the Fraudulent Credit Card Transactions Dataset

Fraud is a major problem for credit card companies, both because of the large volume of transactions that are completed each day and because many fraudulent transactions look a lot like normal transactions. Identifying fraudulent credit card transactions is a common type of imbalanced binary classification where the focus is on the positive class (is […]

Continue Reading
What is Deep Learning?

What is Deep Learning?

A lot is happening in the world of AI at the moment. Some of you may be wondering how machines have the ability to do what they can do. How can they recognise images, understand speech, and even reply to my requests??? Welcome to the world of Deep Learning.  Deep Learning is a subfield of […]

Continue Reading
8 Tricks for Configuring Backpropagation to Train Better Neural Networks, Faster

8 Tricks for Configuring Backpropagation to Train Better Neural Networks

Neural network models are trained using stochastic gradient descent and model weights are updated using the backpropagation algorithm. The optimization solved by training a neural network model is very challenging and although these algorithms are widely used because they perform so well in practice, there are no guarantees that they will converge to a good […]

Continue Reading
Computational Linear Algebra for Coders Review

Computational Linear Algebra for Coders Review

Numerical linear algebra is concerned with the practical implications of implementing and executing matrix operations in computers with real data. It is an area that requires some previous experience of linear algebra and is focused on both the performance and precision of the operations. The company fast.ai released a free course titled “Computational Linear Algebra” […]

Continue Reading
Linear Algebra for Deep Learning

Linear Algebra for Deep Learning

Linear algebra is a field of applied mathematics that is a prerequisite to reading and understanding the formal description of deep learning methods, such as in papers and textbooks. Generally, an understanding of linear algebra (or parts thereof) is presented as a prerequisite for machine learning. Although important, this area of mathematics is seldom covered […]

Continue Reading