Catalog / Machine Learning Cheat Sheet
Machine Learning Cheat Sheet
A comprehensive cheat sheet covering fundamental machine learning concepts, algorithms, and techniques. Useful for quick reference and understanding key aspects of machine learning workflows.
Fundamentals
Key Concepts
Supervised Learning: Learning from labeled data to predict outcomes or classify new data points. Unsupervised Learning: Discovering patterns and structures in unlabeled data. Reinforcement Learning: Training an agent to make decisions in an environment to maximize a reward. |
Model: A mathematical representation of a real-world process, used for prediction or understanding. Features: Input variables used by a model to make predictions. Labels: Output variables that a supervised learning model is trained to predict. |
Training Set: Data used to train a machine learning model. Validation Set: Data used to tune the hyperparameters of a model. Test Set: Data used to evaluate the performance of a trained model. |
Bias-Variance Tradeoff
Bias |
The error due to oversimplification in a learning algorithm. High bias can cause an algorithm to miss relevant relations between features and target outputs (underfitting). |
Variance |
The error due to the model’s sensitivity to small fluctuations in the training data. High variance can cause an algorithm to model the random noise in the training data, rather than the intended outputs (overfitting). |
Tradeoff |
Finding the right balance between bias and variance is crucial for building accurate models. Reducing one often increases the other. |
Common Evaluation Metrics
Accuracy: The proportion of correctly classified instances. Precision: The proportion of true positives among the instances predicted as positive. Recall: The proportion of true positives among the actual positive instances. |
F1-score: The harmonic mean of precision and recall. AUC-ROC: Area under the Receiver Operating Characteristic curve, which measures the ability of a classifier to distinguish between classes. Mean Squared Error (MSE): The average squared difference between predicted and actual values (for regression). |
Supervised Learning
Regression Algorithms
Linear Regression |
Models the relationship between variables using a linear equation. Sensitive to outliers. Assumes linearity. |
Polynomial Regression |
Extends linear regression by adding polynomial terms. Can model non-linear relationships. |
Ridge Regression |
Adds L2 regularization to linear regression to prevent overfitting. |
Lasso Regression |
Adds L1 regularization to linear regression, which can also perform feature selection. |
Support Vector Regression (SVR) |
Uses support vector machines for regression tasks. Effective in high dimensional spaces. |
Classification Algorithms
Logistic Regression |
Predicts the probability of a binary outcome. Uses a sigmoid function. Output between 0 and 1. |
Support Vector Machine (SVM) |
Finds the optimal hyperplane to separate data points into different classes. Effective in high-dimensional spaces. |
Decision Tree |
Uses a tree-like structure to make decisions based on features. Prone to overfitting. |
Random Forest |
An ensemble of decision trees, which reduces overfitting and improves accuracy. |
Naive Bayes |
Applies Bayes’ theorem with strong (naive) independence assumptions between features. Fast and simple. |
K-Nearest Neighbors (KNN) |
Classifies data points based on the majority class among its k nearest neighbors. Sensitive to feature scaling. |
Ensemble Methods
Bagging: Training multiple models on different subsets of the training data and averaging their predictions. Reduces variance. Boosting: Training models sequentially, where each model tries to correct the errors of its predecessors. Reduces bias. |
Examples of Bagging: Random Forest. Examples of Boosting: AdaBoost, Gradient Boosting, XGBoost, LightGBM, CatBoost. |
Unsupervised Learning
Clustering Algorithms
K-Means |
Partitions data into k clusters, where each data point belongs to the cluster with the nearest mean (centroid). Requires specifying the number of clusters (k). |
Hierarchical Clustering |
Builds a hierarchy of clusters, either agglomerative (bottom-up) or divisive (top-down). Doesn’t require specifying the number of clusters beforehand. |
DBSCAN (Density-Based Spatial Clustering of Applications with Noise) |
Groups together data points that are closely packed together, marking as outliers points that lie alone in low-density regions. Does not require specifying the number of clusters and is robust to outliers. |
Spectral Clustering |
Uses the eigenvalues of the similarity matrix of the data to reduce the dimensionality of the data before clustering in fewer dimensions. |
Dimensionality Reduction
Principal Component Analysis (PCA) |
Transforms data into a new coordinate system where the principal components capture the most variance in the data. Reduces the number of features while preserving important information. |
t-distributed Stochastic Neighbor Embedding (t-SNE) |
Reduces the dimensionality of data while preserving the local structure of the data points. Useful for visualizing high-dimensional data in lower dimensions. |
Autoencoders |
Neural networks that learn to encode and decode data, typically used for dimensionality reduction or feature learning. Can capture non-linear relationships. |
Association Rule Mining
Apriori Algorithm: An algorithm for frequent itemset mining and association rule learning over relational databases. Identifies frequent itemsets and generates association rules based on minimum support and confidence thresholds. Eclat Algorithm: Uses a depth-first search approach to find frequent itemsets. Can be more efficient than Apriori for certain datasets. |
Reinforcement Learning
Core Concepts
Agent: An entity that interacts with an environment to learn optimal behavior. Environment: The world in which the agent operates. State: A representation of the current situation of the environment. |
Action: A choice made by the agent to interact with the environment. Reward: A scalar value that measures the immediate feedback from the environment to the agent. Policy: A strategy that the agent uses to determine which action to take in a given state. |
Value Function: A function that estimates the expected cumulative reward from a given state following a particular policy. Q-Function: A function that estimates the expected cumulative reward from taking a specific action in a given state and following a particular policy. |
Algorithms
Q-Learning |
An off-policy algorithm that learns the optimal Q-function by iteratively updating Q-values based on observed rewards. Uses the Bellman equation. |
SARSA (State-Action-Reward-State-Action) |
An on-policy algorithm that updates Q-values based on the current policy. More conservative than Q-learning. |
Deep Q-Network (DQN) |
Uses a deep neural network to approximate the Q-function. Combines Q-learning with deep learning to handle high-dimensional state spaces. |
Policy Gradients |
Directly optimizes the policy without using a value function. Suitable for continuous action spaces. |
Actor-Critic Methods |
Combines policy gradients with value-based methods. Uses an actor to learn the policy and a critic to estimate the value function. |
Exploration vs. Exploitation
Exploration: Trying out different actions to discover new states and rewards. Exploitation: Choosing the action that is believed to yield the highest reward based on current knowledge. Tradeoff: Balancing exploration and exploitation is crucial for learning optimal policies. Common strategies include ε-greedy and Upper Confidence Bound (UCB). |