Главная
Study mode:
on
1
Intro
2
Foundations of Machine Learning
3
The Transformer Architecture
4
Transformer Decoder Overview
5
Inputs
6
Input Embedding
7
Masked Multi-Head Attention
8
Positional Encoding
9
Skip Connections and Layer Norm
10
Feed-forward Layer
11
Transformer hyperparameters and Why they work so well
12
Notable LLM: BERT
13
Notable LLM: T5
14
Notable LLM: GPT
15
Notable LLM: Chinchilla and Scaling Laws
16
Notable LLM: LLaMA
17
Why include code in LLM training data?
18
Instruction Tuning
19
Notable LLM: RETRO
Description:
Dive into a comprehensive 48-minute video lecture on the foundational concepts of large language models. Explore core machine learning principles, the Transformer architecture, and notable LLMs. Gain insights into pretraining dataset composition, including the importance of code in training data. Learn about key components such as input embedding, masked multi-head attention, positional encoding, and feed-forward layers. Discover why Transformers work so well and examine notable models like BERT, T5, GPT, Chinchilla, LLaMA, and RETRO. Understand the significance of scaling laws and instruction tuning in LLM development. Access accompanying slides and additional resources for a deeper understanding of this rapidly evolving field.

LLM Foundations - LLM Bootcamp

The Full Stack
Add to list
0:00 / 0:00