Useful Links
Computer Science
Artificial Intelligence
Machine Learning
Supervised Learning
1. Foundations of Supervised Learning
2. The Supervised Learning Workflow
3. Linear Models
4. Tree-Based Models
5. Instance-Based Learning
6. Support Vector Machines
7. Probabilistic Models
8. Model Evaluation and Validation
9. Advanced Topics in Supervised Learning
10. Practical Implementation Considerations
Tree-Based Models
Decision Trees
Decision Tree Structure
Nodes and Leaves
Decision Rules
Tree Depth and Width
Decision Tree for Classification
Splitting Criteria
Gini Impurity
Mathematical Definition
Calculation Examples
Information Gain and Entropy
Entropy Definition
Information Gain Calculation
Chi-square Test
Tree Construction Algorithm
Recursive Partitioning
Stopping Criteria
Handling Categorical Variables
Feature Importance Calculation
Decision Tree for Regression
Splitting Criteria for Regression
Variance Reduction
Mean Squared Error Reduction
Prediction in Regression Trees
Leaf Value Determination
Tree Pruning
Pre-pruning Strategies
Maximum Depth
Minimum Samples per Leaf
Minimum Impurity Decrease
Post-pruning Strategies
Cost Complexity Pruning
Reduced Error Pruning
Advantages and Limitations
Interpretability Benefits
Handling Non-linear Relationships
Overfitting Tendency
Instability Issues
Random Forest
Ensemble Concept
Bootstrap Aggregating
Bootstrap Sampling
Aggregation Strategies
Random Feature Selection
Feature Randomness at Each Split
Number of Features to Consider
Forest Construction
Building Individual Trees
Combining Predictions
Majority Voting for Classification
Averaging for Regression
Out-of-Bag Evaluation
OOB Error Estimation
Variable Importance
Hyperparameters
Number of Trees
Tree Depth
Feature Subset Size
Advantages over Single Trees
Reduced Overfitting
Improved Generalization
Robustness to Outliers
Gradient Boosting Trees
Boosting Concept
Sequential Learning Process
Gradient Boosting Algorithm
Additive Model Building
Gradient Computation
Weak Learner Training
Loss Functions in Gradient Boosting
Regression Loss Functions
Classification Loss Functions
Regularization Techniques
Learning Rate
Tree Constraints
Subsampling
Popular Implementations
XGBoost
Regularization Features
Handling Missing Values
Parallel Processing
LightGBM
Leaf-wise Tree Growth
Gradient-based One-Side Sampling
Exclusive Feature Bundling
CatBoost
Categorical Feature Handling
Ordered Boosting
Symmetric Trees
Previous
3. Linear Models
Go to top
Next
5. Instance-Based Learning