Useful Links
Statistics
Machine Learning
1. Introduction to Machine Learning
2. Mathematical and Statistical Foundations
3. Data Preprocessing and Feature Engineering
4. Supervised Learning
5. Unsupervised Learning
6. Model Evaluation and Validation
7. Ensemble Methods and Advanced Techniques
8. Deep Learning and Neural Networks
9. Reinforcement Learning
10. Advanced Topics and Specialized Areas
11. Machine Learning Operations and Deployment
Supervised Learning
Regression Analysis
Linear Regression
Simple Linear Regression
Model Assumptions
Least Squares Method
Coefficient Interpretation
Multiple Linear Regression
Multiple Predictors
Coefficient Interpretation
Multicollinearity
Model Assumptions
Linearity
Independence
Homoscedasticity
Normality of Residuals
Cost Function
Mean Squared Error
Sum of Squared Residuals
Optimization Methods
Analytical Solution
Gradient Descent
Stochastic Gradient Descent
Model Evaluation
Residual Analysis
R-Squared
Adjusted R-Squared
F-Statistic
Polynomial Regression
Polynomial Features
Model Complexity
Overfitting Issues
Cross-Validation for Model Selection
Regularization Techniques
Ridge Regression
L2 Penalty
Bias-Variance Tradeoff
Hyperparameter Tuning
Lasso Regression
L1 Penalty
Feature Selection Property
Sparse Solutions
Elastic Net
Combined L1 and L2 Penalties
Mixing Parameter
Regularization Path
Cross-Validation for Regularization
Non-Linear Regression
Basis Functions
Kernel Methods
Spline Regression
Robust Regression
Outlier-Resistant Methods
Huber Loss
Quantile Regression
Classification Methods
Logistic Regression
Binary Classification
Sigmoid Function
Odds and Log-Odds
Maximum Likelihood Estimation
Multinomial Classification
Softmax Function
One-vs-Rest
One-vs-One
Cost Function
Cross-Entropy Loss
Log-Likelihood
Optimization
Gradient Descent
Newton-Raphson Method
Model Interpretation
Coefficient Interpretation
Odds Ratios
Regularization in Logistic Regression
L1 Regularization
L2 Regularization
k-Nearest Neighbors
Algorithm Mechanics
Distance Metrics
Euclidean Distance
Manhattan Distance
Minkowski Distance
Hamming Distance
Choosing k
Odd vs. Even k
Cross-Validation for k Selection
Weighted k-NN
Computational Complexity
Curse of Dimensionality
Efficient Implementations
KD-Trees
Ball Trees
LSH
Support Vector Machines
Linear SVM
Maximum Margin Principle
Support Vectors
Hard Margin Classification
Soft Margin Classification
Slack Variables
Non-Linear SVM
Kernel Trick
Kernel Functions
Linear Kernel
Polynomial Kernel
RBF Kernel
Sigmoid Kernel
Kernel Selection
SVM for Regression
Support Vector Regression
Epsilon-Insensitive Loss
Hyperparameter Tuning
C Parameter
Gamma Parameter
Kernel Parameters
Computational Considerations
Quadratic Programming
SMO Algorithm
Naive Bayes Classifiers
Bayes' Theorem in Classification
Naive Independence Assumption
Types of Naive Bayes
Gaussian Naive Bayes
Multinomial Naive Bayes
Bernoulli Naive Bayes
Complement Naive Bayes
Laplace Smoothing
Advantages and Limitations
Text Classification Applications
Decision Trees
Tree Structure
Root Node
Internal Nodes
Leaf Nodes
Branches
Splitting Criteria
Information Gain
Entropy
Gini Impurity
Gain Ratio
Chi-Square
Tree Construction Algorithms
ID3
C4.5
CART
Handling Different Data Types
Categorical Features
Numerical Features
Missing Values
Pruning Techniques
Pre-Pruning
Post-Pruning
Cost Complexity Pruning
Advantages and Limitations
Interpretability
Non-Linear Relationships
Overfitting Tendency
Linear Discriminant Analysis
Fisher's Linear Discriminant
Assumptions
Dimensionality Reduction
Comparison with PCA
Quadratic Discriminant Analysis
Relaxed Assumptions
Quadratic Decision Boundaries
Previous
3. Data Preprocessing and Feature Engineering
Go to top
Next
5. Unsupervised Learning