Главная
Study mode:
on
1
intro
2
starter code
3
fixing the initial loss
4
fixing the saturated tanh
5
calculating the init scale: “Kaiming init”
6
batch normalization
7
batch normalization: summary
8
real example: resnet50 walkthrough
9
summary of the lecture
10
just kidding: part2: PyTorch-ifying the code
11
viz #1: forward pass activations statistics
12
viz #2: backward pass gradient statistics
13
the fully linear case of no non-linearities
14
viz #3: parameter activation and gradient statistics
15
viz #4: update:data ratio over time
16
bringing back batchnorm, looking at the visualizations
17
summary of the lecture for real this time
Description:
Dive deep into the internals of multi-layer perceptrons (MLPs) in this comprehensive video lecture. Explore the statistics of forward pass activations and backward pass gradients, while learning about potential pitfalls in improperly scaled networks. Discover essential diagnostic tools and visualizations for assessing the health of deep networks. Understand the challenges of training deep neural networks and learn about Batch Normalization, a key innovation that simplifies the process. Gain practical insights through code examples, real-world applications, and visualizations. Complete provided exercises to reinforce your understanding of weight initialization and BatchNorm implementation. Follow along as the lecture covers topics such as Kaiming initialization, PyTorch implementation, and various visualization techniques for network analysis.

Building Makemore - Activations & Gradients, BatchNorm

Andrej Karpathy
Add to list
0:00 / 0:00