Useful Links
Computer Science
Artificial Intelligence
Deep Learning
Reinforcement Learning
1. Foundations of Reinforcement Learning
2. Mathematical Foundations
3. Markov Decision Processes
4. Dynamic Programming
5. Monte Carlo Methods
6. Temporal-Difference Learning
7. Function Approximation
8. Deep Reinforcement Learning
9. Policy Gradient Methods
10. Advanced Topics
11. Implementation and Practical Considerations
12. Applications and Case Studies
Function Approximation
Need for Function Approximation
Large State Spaces
Continuous State Spaces
Curse of Dimensionality
Generalization Requirements
Value Function Approximation
Approximate Value Functions
Function Approximation Architectures
Feature Representation
Hand-Crafted Features
Basis Functions
Feature Selection
Linear Function Approximation
Linear Combinations of Features
Weight Vector Learning
Convergence Properties
Nonlinear Function Approximation
Neural Networks
Decision Trees
Kernel Methods
Prediction with Function Approximation
Gradient-Based Methods
Stochastic Gradient Descent
Learning Rate Schedules
Convergence Analysis
Semi-Gradient Methods
Bootstrapping with Function Approximation
Stability Issues
Convergence Conditions
Least-Squares Methods
LSTD (Least-Squares TD)
Computational Complexity
Batch Updates
Control with Function Approximation
Action-Value Function Approximation
Semi-Gradient SARSA
Semi-Gradient Q-Learning
Policy Gradient Methods
Stability and Convergence Issues
The Deadly Triad
Function Approximation
Bootstrapping
Off-Policy Learning
Divergence Examples
Stabilization Techniques
Feature Construction
Tile Coding
Radial Basis Functions
Fourier Basis
Polynomial Features
Previous
6. Temporal-Difference Learning
Go to top
Next
8. Deep Reinforcement Learning