Useful Links
Computer Science
Artificial Intelligence
Deep Learning
Deep Learning and Neural Networks
1. Foundations of Machine Learning and Neural Networks
2. Training Shallow Neural Networks
3. Deepening the Network
4. Practical Considerations for Training
5. Convolutional Neural Networks (CNNs)
6. Recurrent Neural Networks (RNNs)
7. The Transformer Architecture
8. Generative Models
9. Deep Reinforcement Learning
10. Advanced Topics and Specialized Architectures
11. Deployment and Production
Practical Considerations for Training
Advanced Optimization Algorithms
Limitations of Basic Gradient Descent
Momentum-Based Methods
Momentum
Concept and Implementation
Effect on Convergence
Momentum Parameter Selection
Nesterov Accelerated Gradient
Adaptive Learning Rate Methods
AdaGrad
Adaptive Learning Rates
Accumulation of Squared Gradients
Strengths and Weaknesses
RMSprop
Running Average of Squared Gradients
Decay Rate Parameter
Use in Practice
Adam Optimizer
Combination of Momentum and RMSprop
Bias Correction
Hyperparameters
AdamW
Nadam
Second-Order Methods
Newton's Method
Quasi-Newton Methods
L-BFGS
Regularization Techniques
Weight Regularization
L1 Regularization (Lasso)
Mathematical Formulation
Sparsity Induction
L2 Regularization (Ridge)
Mathematical Formulation
Weight Decay Effect
Elastic Net Regularization
Dropout
Random Neuron Deactivation
Training vs. Inference Behavior
Dropout Rate Selection
Variants of Dropout
Early Stopping
Monitoring Validation Loss
Patience Parameter
Stopping Criteria
Data Augmentation
Techniques for Images
Techniques for Text Data
Techniques for Tabular Data
Synthetic Data Generation
Ensemble Methods
Model Averaging
Bagging
Boosting
Hyperparameter Tuning
Importance of Hyperparameters
Hyperparameter Categories
Architecture Hyperparameters
Training Hyperparameters
Regularization Hyperparameters
Search Strategies
Grid Search
Exhaustive Search Strategy
Computational Complexity
Random Search
Random Sampling of Hyperparameters
Efficiency Advantages
Bayesian Optimization
Probabilistic Model-Based Search
Acquisition Functions
Evolutionary Algorithms
Learning Rate Scheduling
Step Decay
Exponential Decay
Cosine Annealing
Cyclical Learning Rates
Warm Restarts
Cross-Validation Strategies
K-Fold Cross-Validation
Stratified Cross-Validation
Time Series Cross-Validation
Previous
3. Deepening the Network
Go to top
Next
5. Convolutional Neural Networks (CNNs)