Confusion Matrix: A table that describes the performance of a classification model.
- Key Metrics: True Positives (TP), True Negatives (TN), False Positives (FP), False Negatives (FN).
✅ 1. Supervised Learning
• Regression
o Linear Regression
o Logistic Regression
o Polynomial Regression
o Ridge Regression
o Lasso Regression
o ElasticNet
o Support Vector Machines (SVM)
o Decision Trees
o Random Forest
• Classification
o Logistic Regression
o K-Nearest Neighbors (KNN)
o Support Vector Machines (SVM)
o Decision Trees
o Random Forest
o Naive Bayes
o Confusion Matrix
o Stochastic Gradient Descent
o Gradient Boosting
o AdaBoost
o XGBoost
o LightGBM
o CatBoost
________________________________________
🔍 2. Unsupervised Learning
• Clustering
🔹 1. Centroid-Based Clustering
• K-Means
• K-Medoids
• Mean-Shift
________________________________________
🔹 2. Density-Based Clustering
• DBSCAN
• OPTICS
• HDBSCAN
________________________________________
🔹 3. Hierarchical Clustering
• Agglomerative Clustering
• BIRCH (Balanced Iterative Reducing and Clustering using Hierarchies)
• Affinity Propagation
________________________________________
🔹 4. Distribution-Based Clustering
• Gaussian Mixture Models (GMM)
• Dimensionality Reduction
o PCA (Principal Component Analysis)
o t-SNE
o UMAP
o ICA (Independent Component Analysis)
o LDA (Linear Discriminant Analysis)
________________________________________
🔁 3. Semi-Supervised Learning
• Self-Training
• Label Propagation
• Label Spreading
________________________________________
🔄 4. Reinforcement Learning
• Q-Learning
• Deep Q-Networks (DQN)
• SARSA
• Policy Gradient Methods
• Actor-Critic
• Proximal Policy Optimization (PPO)
• Deep Deterministic Policy Gradient (DDPG)
________________________________________
🧠 5. Deep Learning Algorithms
🔹 1. Feedforward Networks (FNN)
• Multilayer Perceptron (MLP)
• Deep Neural Networks (DNN)
________________________________________
🔹 2. Convolutional Neural Networks (CNN)
• LeNet
• AlexNet
• VGGNet
• GoogLeNet (Inception)
• ResNet
• DenseNet
• EfficientNet
• MobileNet
• SqueezeNet
________________________________________
🔹 3. Recurrent Neural Networks (RNN)
• Vanilla RNN
• Long Short-Term Memory (LSTM)
• Gated Recurrent Unit (GRU)
• Bidirectional RNN
• Deep RNNs
• Echo State Networks (ESN)
________________________________________
🔹 4. Attention-Based Models / Transformers
• Transformer
• BERT
• GPT (GPT-1, GPT-2, GPT-3, GPT-4)
• RoBERTa
• ALBERT
• XLNet
• T5
• DistilBERT
• Vision Transformer (ViT)
• Swin Transformer
• DeiT
• Performer
• Longformer
________________________________________
🔹 5. Autoencoders
• Vanilla Autoencoder
• Sparse Autoencoder
• Denoising Autoencoder
• Contractive Autoencoder
• Variational Autoencoder (VAE)
________________________________________
🔹 6. Generative Adversarial Networks (GANs)
• Vanilla GAN
• Deep Convolutional GAN (DCGAN)
• Conditional GAN (cGAN)
• CycleGAN
• StyleGAN
• Pix2Pix
• BigGAN
• StarGAN
• WGAN (Wasserstein GAN)
• WGAN-GP
________________________________________
🔹 7. Reinforcement Learning (Deep RL)
• Deep Q-Network (DQN)
• Double DQN
• Dueling DQN
• Policy Gradient
• REINFORCE
• Actor-Critic
• A3C (Asynchronous Advantage Actor-Critic)
• PPO (Proximal Policy Optimization)
• DDPG (Deep Deterministic Policy Gradient)
• TD3 (Twin Delayed DDPG)
• SAC (Soft Actor-Critic)
Confusion Matrix: A table that describes the performance of a classification model.
|
Accuracy: (TP + TN) / Total
|
Precision: TP / (TP + FP)
|
Recall (Sensitivity): TP / (TP + FN)
|
F1-Score: 2 * (Precision * Recall) / (Precision + Recall)
|
AUC-ROC: Area Under the Receiver Operating Characteristic curve.
|
Linear Regression: Models the relationship between variables by fitting a linear equation to observed data.
|
Logistic Regression: Predicts the probability of a categorical outcome.
|
Polynomial Regression: Models non-linear relationships using polynomial functions.
|
Ridge Regression: Linear regression with L2 regularization to prevent overfitting.
|
Lasso Regression: Linear regression with L1 regularization, performs feature selection.
|
Elastic Net: Combines L1 and L2 regularization.
|
Support Vector Regression (SVR),Random Forest,Decision Trees |
Logistic Regression: (Also used for classification) Predicts probability of a binary outcome.
|
K-Nearest Neighbors (KNN): Classifies based on the majority class among its k nearest neighbors.
|
Support Vector Machines (SVM): Finds an optimal hyperplane to separate classes.
|
Decision Trees: (Also used for classification) Splits data into subsets based on feature values.
|
Random Forest: (Also used for classification) Ensemble of decision trees for improved accuracy.
|
Naive Bayes: Applies Bayes’ theorem with strong (naive) independence assumptions between features.
|
Stochastic Gradient Descent (SGD): Optimization algorithm used to train linear classifiers under convex loss functions.
|
Gradient Boosting: Builds an ensemble of weak learners sequentially, where each learner corrects the errors of its predecessors.
|
AdaBoost: Adaptive Boosting, focuses on correcting mistakes of previous classifiers.
|
XGBoost: Optimized Gradient Boosting implementation.
|
LightGBM: Gradient Boosting framework that uses tree based learning algorithms.
|
CatBoost: Gradient Boosting algorithm that handles categorical features natively.
|
K-Means: Partitions n observations into k clusters, each with the nearest mean.
|
K-Medoids: Similar to K-Means but chooses data points as cluster centers (medoids).
|
Mean-Shift: Locates the maxima of a density function.
|
DBSCAN: Density-Based Spatial Clustering of Applications with Noise; identifies clusters based on density.
|
OPTICS: Ordering Points To Identify the Clustering Structure; extends DBSCAN to handle varying densities.
|
HDBSCAN: Hierarchical DBSCAN; combines hierarchical clustering with DBSCAN.
|
Agglomerative Clustering: Bottom-up approach where each observation starts in its own cluster, and pairs of clusters are merged as one moves up the hierarchy.
|
BIRCH: Balanced Iterative Reducing and Clustering using Hierarchies; designed for large datasets.
|
Affinity Propagation: Creates clusters by sending messages between pairs of samples until convergence.
|
Gaussian Mixture Models (GMM): Assumes data is generated from a mixture of Gaussian distributions.
|
PCA (Principal Component Analysis): Reduces dimensionality by projecting data onto principal components.
|
t-SNE: Reduces dimensionality while keeping similar instances close and dissimilar instances apart.
|
UMAP: Uniform Manifold Approximation and Projection; similar to t-SNE but faster and can preserve more global structure.
|
ICA (Independent Component Analysis): Separates multivariate signals into additive subcomponents that are statistically independent.
|
LDA (Linear Discriminant Analysis): Finds a linear combination of features that characterizes or separates two or more classes of objects or events.
|
Self-Training: Train a model on labeled data, predict labels for unlabeled data, add high-confidence predictions to the labeled set, and retrain.
|
Label Propagation: Assigns labels to unlabeled data points based on the labels of their neighbors.
|
Label Spreading: Similar to label propagation but uses a graph-based approach to spread labels.
|
Q-Learning: Model-free RL algorithm that learns a Q-function representing the expected reward for taking an action in a state.
|
Deep Q-Networks (DQN): Uses a neural network to approximate the Q-function.
|
SARSA: On-policy RL algorithm that updates the Q-function based on the action actually taken.
|
Policy Gradient Methods: Directly optimize the policy function.
|
Actor-Critic: Combines policy gradient and value-based methods.
|
Proximal Policy Optimization (PPO): Policy gradient method that constrains policy updates.
|
Deep Deterministic Policy Gradient (DDPG): Actor-critic method that handles continuous action spaces.
|
LeNet: Early CNN architecture for handwritten digit recognition.
|
AlexNet: Deeper CNN architecture that won the 2012 ImageNet competition.
|
VGGNet: CNN architecture with very deep layers and small convolutional filters.
|
GoogLeNet (Inception): Uses inception modules to reduce computational cost and improve performance.
|
ResNet: Introduces residual connections to train very deep networks.
|
DenseNet: Connects each layer to every other layer in a feed-forward fashion.
|
EfficientNet: Balances network depth, width, and resolution.
|
MobileNet: Designed for mobile and embedded devices.
|
SqueezeNet: Achieves AlexNet-level accuracy with fewer parameters.
|
Deep Neural Networks (DNN): Neural networks with multiple hidden layers.
|
Multilayer Perceptron (MLP): Basic feedforward neural network with one or more hidden layers.
|
Vanilla RNN: Basic RNN architecture for sequential data.
|
Long Short-Term Memory (LSTM): RNN architecture with memory cells to capture long-range dependencies.
|
Gated Recurrent Unit (GRU): Simplified version of LSTM.
|
Bidirectional RNN: Processes sequences in both forward and backward directions.
|
Deep RNNs: RNNs with multiple layers.
|
Echo State Networks (ESN): RNN with a sparsely connected hidden layer.
|