Главная
Study mode:
on
1
Transformers are taking over AI right now, and quite possibly their most famous use is in ChatGPT. ChatGPT uses a specific type of Transformer called a Decoder-Only Transformer, and this StatQuest sh…
2
Awesome song and introduction
3
Word Embedding
4
Position Encoding
5
Masked Self-Attention, an Autoregressive method
6
Residual Connections
7
Generating the next word in the prompt
8
Review of encoding and generating the prompt
9
Generating the output, Part 1
10
Masked Self-Attention while generating the output
11
Generating the output, Part 2
12
Normal Transformers vs Decoder-Only Transformers
Description:
Dive into a comprehensive 37-minute video tutorial exploring Decoder-Only Transformers, the specific type of Transformer used in ChatGPT. Learn about word embedding, position encoding, masked self-attention as an autoregressive method, and residual connections. Understand the process of generating the next word in a prompt, encoding and generating prompts, and the two-part output generation process. Compare Normal Transformers with Decoder-Only Transformers, and gain insights into the inner workings of cutting-edge AI technology. Supplementary resources for deeper understanding of related concepts like backpropagation, SoftMax function, and word embedding are also provided.

Decoder-Only Transformers, ChatGPT's Specific Transformer, Clearly Explained

StatQuest with Josh Starmer
Add to list
0:00 / 0:00