Главная
Study mode:
on
1
– Welcome to class
2
– Listening to YouTube from the terminal
3
– Summarising papers with @Notion
4
– Reading papers collaboratively
5
– Attention! Self / cross, hard / soft
6
– Use cases: set encoding!
7
– Self-attention
8
– Key-value store
9
– Queries, keys, and values → self-attention
10
– Queries, keys, and values → cross-attention
11
– Implementation details
12
– The Transformer: an encoder-predictor-decoder architecture
13
– The Transformer encoder
14
– The Transformer “decoder” which is an encoder-predictor-decoder module
15
– Jupyter Notebook and PyTorch implementation of a Transformer encoder
16
– Goodbye :
Description:
Explore the intricacies of attention mechanisms and Transformer architecture in this comprehensive lecture. Delve into self-attention, cross-attention, hard attention, and soft attention concepts. Learn about set encoding use cases and the key-value store paradigm. Understand the implementation of queries, keys, and values in both self-attention and cross-attention contexts. Examine the Transformer's encoder-predictor-decoder architecture, with a focus on the encoder and the unique "decoder" module. Gain practical insights through a PyTorch implementation of a Transformer encoder using Jupyter Notebook. Additionally, discover useful tips for reading and summarizing research papers collaboratively.

Self - Cross, Hard - Soft Attention and the Transformer

Alfredo Canziani
Add to list
0:00 / 0:00