Unlocking the Power of Machine Learning: A Beginner’s Guide to Understanding Algorithms and Models

Mahesh Jadhav
11 min readApr 18, 2023

--

Oh man, learning machine learning can be quite an adventure! As I dive into its depths, I realized just how complex and confusing it can be. With the plethora of models and algorithms involved, understanding each one can feel like a never-ending journey. And don’t even get me started on the challenge of perfectly understanding each individual algorithm, as soon as I began to understand a new algorithm, I found myself forgetting the earlier ones. It was like trying to figure out the timeline of the Marvel Cinematic Universe. So, I asked myself, “What can I do to make sense of this madness?” And that’s when I had an Idea.

I decided to take an eagle’s-eye view of the situation and created a map of algorithms depicting machine learning, complete with all its subsets and famous models. This gave me a broader understanding of which algorithm belongs to what model. I also wrote down the description of each algorithm as briefly as possible, giving me an instant idea of its purpose.

This approach actually worked! It helped me to understand and grasp the machine learning models more easily, and I’m confident it can do the same for you too. So, take a deep breath, put on your earbuds, and let’s tackle machine learning together!

Machine learning is a branch of Artificial Intelligence (AI) and computer science that involves statistical and mathematical approaches to find patterns and insights into the data which is then used for building algorithms and models that can learn and make predictions without external instructions. There are many machine learning algorithms that can be broadly classified into 4 types Supervised learning, Unsupervised Learning, Semi-Supervised learning, and Reinforcement learning as shown below.

Types of Machine Learning Models

Supervised Machine Learning

It is a type of machine learning model where algorithms are trained on labeled data to learn the relationships between feature matrix and target variable so that it can make accurate predictions on new, unseen data.

Regression Model

It provides a set of statistical processes to describe the relationship between independent variables and dependent (target) variables.

  • Linear Regression: It is used to predict the value of a continuous dependent variable based on one or more independent variables with the help of a best fit linear line that minimizes the difference between predicted values and the actual values.
  • Logistic Regression: It uses a sigmoid function to estimate the probability of an event occurring based on a given dataset of independent variables. It is primarily used for binary classification problems where the output variable can only take two values (0 | 1).

Classification Model

Supervised machine learning method where algorithm tries to predict or classify data into predefined categories based on characteristics.

  • Decision Tree: A specific type of probability tree where each node represents a Yes|No type of question and the branch represents outcome of test that enables the algorithm to make decisions.
  • Random Forest: It creates multiple decision trees using a random subset of training data and a random subset of features, which are then combined to make a final prediction.
  • KNN (K-Nearest Neighbor): It is a non-parametric supervised learning method which calculates the K closest neighbors of a given data point in the training set, and then identifies the majority class among those neighbors as the predicted class for that data point. K’s value is a hyperparameter that can be adjusted to improve algorithm’s efficiency.
  • SVM (Support Vector Machine): SVM tries to separate data by drawing a plane between them. We choose the plane in such a way that it creates the largest possible gap between the two types of data. This gap is called the margin. The points closest to the plane are called support vectors. By using these support vectors, we can classify new data into one of the two groups.
  • Naive Bayes Classification: It predicts outcomes based on a set of input features. It uses probability theory to determine the likelihood of an outcome based on the presence or absence of certain features. The “naive” part of its name comes from the assumption that each feature is independent of the others, which simplifies the calculations.

Unsupervised Machine Learning

Type of Machine learning technique where the algorithm learns to recognize patterns in the data without the use of labeled examples or external guidance.

Clustering Models

These are machine learning models used to group similar data points together based on their features. They don’t rely on labeled data to make predictions. Instead, they try to find patterns and relationships in the data on their own.

  • K - Means Clustering: This algorithm partitions the dataset into K-Clusters where K is a predefined number. Its goal is to optimize the sum of squared distance between data points and their assigned cluster centroid.
  • Hierarchical Model: This method starts by considering each item as its own cluster, and then repeatedly merges the closest pairs of clusters until all items are in a single group. The resulting tree can be cut at different levels to obtain different levels of clustering. It is useful for exploring relationships between items in a dataset and identifying natural groupings using a tree-like structure called a dendrogram.
  • DBSCAN: Density-based spatial clustering of applications with noise does not require the number of clusters to be specified beforehand. Instead, it determines clusters based on the density of points in the data set. Points that are close together and have high density are considered part of the same cluster, while points that are isolated or have low density are classified as noise.
  • GMM (Gaussian Mixture Model): It assumes that the data comes from a mixture of many Gaussian distributions. Each of these distributions represents a different group or cluster in the data. The model tries to estimate the parameters of these distributions (such as mean and variance) and the probability of each data point belonging to each distribution
  • Spectral Clustering Model: It creates a similarity graph of the data set, where each data point is represented as a node in the graph and the edges between the nodes represent their pairwise similarity. Spectral clustering then uses the graph’s spectral properties to identify clusters in the data by finding the eigenvectors (principal components) of the graph’s Laplacian matrix to project data into lower-dimensional space, where it is easier to identify clusters.
  • Mean-Shift Clustering: It works by finding the centroids or mean points of each cluster by shifting a window around the data space to the areas of higher density. The window shifts towards the steepest ascent until it reaches a peak where the density is highest, and this peak is then designated as the centroid of the cluster. The window is again moved to a new peak and this process is repeated until all centroids are found.
  • SOM (Self-Organizing Map): It is a type of artificial neural network that works by mapping a high-dimensional data set onto a low-dimensional grid of neurons in a way that preserves the topological relationship between the data points. Each neuron in the grid represents a cluster or a group of similar data points.

Dimensionality Reduction Models

These are techniques in machine learning that simplify complex data sets by reducing the number of features or variables while retaining the most important information. They help to remove noise and redundant information from the data, making it easier to analyze and visualize.

  • PCA (Principal Component Analysis): It transforms the data set into a new coordinate system, where the new axes represent principal components that capture the maximum variance in the data. These principal components are linear combinations of the original features, and they are ordered by the amount of variance they explain in the data. By projecting the data onto the first few principal components, PCA can reduce the dimensionality while retaining most important information.
  • NMF (Non-negative Matrix Factorization): It works by factorizing a non-negative matrix (V) into two lower-rank non-negative matrices W and H, such that V=WH. Here, V is a data matrix with rows representing samples and columns representing features, W is a basis matrix with rows representing basis vectors, and H is a coefficient matrix with columns representing coefficients.
  • LDA (Linear Discriminant Analysis): It works by finding a linear combination of features that maximizes the separation between different classes or categories in the data. By projecting the data onto this new linear subspace, LDA can reduce the dimensionality of the data while retaining the most important information for classification.
  • ICA (Independent Component Analysis): It finds a set of independent components that represent the underlying sources of variation in a complex data set. By separating the data into its independent components, ICA can reduce the dimensionality of the data and extract the most important features or signals.
  • Autoencoder: It works by compressing the input data into a lower-dimensional representation, and then reconstructing it back to its original form. By minimizing the difference between the input and output data, the autoencoder can learn a compact and informative representation of the input data.

Semi-Supervised Learning

A machine learning approach where a model is trained on a combination of labeled and unlabeled data. In this approach, the labeled data is used to teach the model to recognize patterns and make predictions, while the unlabeled data is used to improve the model’s generalization and robustness.

  • Self-Training Model: The model is trained on a small set of labeled data, and then used to predict the labels of a larger set of unlabeled data. The predicted labels are then added to the labeled dataset, and the model is retrained on the expanded dataset. This process is repeated iteratively, with the model learning from its own predictions and gradually improving its accuracy on the unlabeled data.
  • Co-Training Model: Two separate models are trained on different subsets of data, one with labeled data and the other with unlabeled data. The models exchange information by using their own predictions as additional labeled data for the other model. This iterative process is repeated, with the models learning from each other’s predictions and gradually improving their accuracy on the unlabeled data.
  • Generative Model: The model learns to identify patterns in the labeled data and then uses those patterns to generate new examples. These generated examples are then used to improve the model’s performance on the unlabeled data.
  • Label Propagation: It learns from both labeled and unlabeled data to make predictions on the unlabeled data. The model first assigns labels to the labeled data and then propagates these labels to the unlabeled data based on the similarity between data points. The model iteratively updates the labels of the unlabeled data until a certain convergence criterion is met.
  • Semi-Supervised Clustering: The model uses the labeled data to learn the initial structure of the clusters and then groups the unlabeled data points based on their similarity to the labeled data points. The model iteratively refines the clustering results using both labeled and unlabeled data until the goal is reached.

Reinforcement Learning

It is a machine learning technique that involves training an agent to make decisions in an environment to maximize a reward. The agent learns through trial and error, receiving feedback in the form of rewards or penalties based on its actions. The goal is for the agent to learn the optimal policy, or sequence of actions, that will result in the highest total reward over time.

Model-Based Learning

It involves learning a model of the environment in addition to learning the optimal policy. The model is a representation of the dynamics of the environment, which the agent can use to predict the next state and reward given the current state and action. The agent can then use these predictions to plan ahead and select the action that will lead to the highest expected reward.

  • MPC (Model Predictive Control): In this algorithm the agent solves an optimization problem at each time step, using the predictive model to simulate future states and rewards. The solution to the optimization problem gives the best sequence of actions to take in the immediate future. MPC is useful when the environment is complex and has a long time horizon, as it allows the agent to plan ahead and make optimal decisions based on predicted outcomes.
  • DP (Dynamic Programming): It involves learning a value function or a policy by solving a system of Bellman equations. It works by iteratively updating the value of each state based on the values of its neighboring states, until convergence. This method is best suited for environments with known dynamics, where the optimal policy can be computed analytically.
  • iLQR (iterative Linear Quadratic Regulator): It is used to solve optimal control problems. It involves iteratively solving a set of linear-quadratic subproblems to find a sequence of control inputs that minimizes a cost function while satisfying a set of constraints. It is often used in robotics and control systems to find optimal control policies for complex systems.
  • Dyna-Q: It is a model-based reinforcement learning algorithm used to find the optimal action policy in a Markov decision process (MDP). It combines model-free and model-based methods to make predictions and learn from experience. It maintains a Q-table to approximate the optimal action-value function and also learns a model of the environment to plan future actions.

Model-Free Learning

technique where an agent learns to make decisions through trial and error without having explicit knowledge of the environment or a model of it. The agent learns by directly interacting with the environment and updating its policy based on the observed rewards.

  • Q-Learning: It enables an agent to learn to make optimal decisions in an environment by finding the best actions to take based on the current state. The algorithm updates its Q-values, which represent the expected reward for taking a particular action in a particular state, through trial and error by exploring the environment and receiving rewards. The agent continues to learn and improve its decision-making skills over time by adjusting its Q-values based on the rewards it receives, and ultimately, it aims to learn the optimal policy for maximizing long-term rewards.
  • SARSA (State-Action-Reward-State-Action): It learns from experiences (trial and error). The agent interacts with the environment, and at each time step, it observes the current state, takes an action, receives a reward, and observes the next state. SARSA algorithm updates its Q-values based on the current observed state, the action taken, the reward received, and the next state and the action that the agent chooses. Unlike Q-Learning, SARSA uses the same policy to select actions during learning and execution, making it an on-policy algorithm.
  • DQN (Deep Q-network): This algorithm combines the Q-learning algorithm with deep neural networks. It is used to solve problems where the state space is too large to be handled by traditional Q-learning. The DQN algorithm works by approximating the Q-value function using a deep neural network and uses it to update the Q-values iteratively. The neural network is trained on the experiences collected from the environment using a technique called experience replay.
  • Monte Carlo Method: It uses repeated random sampling to estimate the value of an action in a given state. It does not require knowledge of the transition probabilities between states or the rewards associated with taking actions. Instead, it learns by randomly exploring the environment and accumulating rewards over time to update the Q-values of each action-state pair.
  • Policy Gradient Method: It involves training a neural network to directly output the policy that maximizes the expected reward. Unlike value-based methods such as Q-learning, policy gradient methods do not estimate the value of each action, but instead optimize the policy directly. The algorithm uses stochastic gradient descent to update the parameters of the policy network to maximize the expected reward.
  • Actor-Critic Method: Type of reinforcement learning where there are two models working together: the actor, which decides on actions to take, and the critic, which evaluates the actions taken by the actor. The actor model determines the best action to take in a given state, and the critic model evaluates how good the chosen action is. By working together, the actor-critic method aims to learn an optimal policy that maximizes rewards over time.

--

--

Mahesh Jadhav
Mahesh Jadhav

Written by Mahesh Jadhav

Part time developer, Full time debugger...

No responses yet