Useful Links
Computer Science
Artificial Intelligence
Natural Language Processing (NLP)
Natural Language Processing (NLP)
1. Introduction to Natural Language Processing
2. Linguistic Foundations
3. Text Processing and Preprocessing
4. Language Modeling
5. Feature Representation
6. Word Embeddings and Distributed Representations
7. Classical Machine Learning for NLP
8. Deep Learning Foundations
9. Recurrent Neural Networks
10. Attention Mechanisms and Transformers
11. Pre-trained Language Models
12. Core NLP Applications
13. Advanced Topics
14. Evaluation and Benchmarking
15. Ethics and Responsible AI
Attention Mechanisms and Transformers
Attention Fundamentals
Motivation and Intuition
Query-Key-Value Framework
Attention Weights and Alignment
Attention Variants
Additive Attention
Multiplicative Attention
Scaled Dot-Product Attention
Multi-Head Attention
Self-Attention
Intra-sequence Dependencies
Positional Information
Computational Complexity
Transformer Architecture
Encoder-Decoder Structure
Positional Encodings
Layer Normalization
Residual Connections
Feed-Forward Networks
Training Transformers
Teacher Forcing
Masked Language Modeling
Autoregressive Generation
Previous
9. Recurrent Neural Networks
Go to top
Next
11. Pre-trained Language Models