Главная
Study mode:
on
1
Intro
2
The Architecture of the Transformer
3
Model Training
4
Transformer LM Component 1: FFNN
5
Transformer LM Component 2: Self-Attention
6
Tokenization: Words to Token Ids
7
Embedding: Breathe meaning into tokens
8
Projecting the Output: Turning Computation into Language
9
Final Note: Visualizing Probabilities
Description:
Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only! Grab it Explore the Transformer architecture, the foundation of state-of-the-art AI/ML models like BERT and GPT, in this 30-minute visual presentation. Delve into the components of Transformer language models, including feed-forward neural networks and self-attention mechanisms. Learn about tokenization, embedding, and output projection processes. Gain insights into model training and probability visualization. Suitable for viewers with various levels of machine learning experience, this accessible video provides a comprehensive overview of the Transformer model's structure and applications in natural language processing.

The Narrated Transformer Language Model

Jay Alammar
Add to list
0:00 / 0:00