Главная
Study mode:
on
1
Loss Landscape and Performance in Deep Learning
2
Supervised Deep Learning
3
Set-up: Architecture
4
Set-up: Dataset
5
Learning
6
Learning dynamics = descent in loss landscape
7
Analogy with granular matter: Jamming
8
Theoretical results: Phase diagram
9
Empirical tests: MNIST parity
10
Landscape curvature
11
Flat directions
12
Outline
13
Overfitting?
14
Ensemble average
15
Fluctuations increase error
16
Scaling argument!
17
Infinitely-wide networks: Initialization
18
Infinitely-wide networks: Learning
19
Neural Tangent Kernel
20
Finite N asymptotics?
21
Conclusion
Description:
Explore the intricacies of deep learning performance and loss landscapes in this 46-minute conference talk. Delve into supervised deep learning concepts, including network architecture and dataset setup. Examine learning dynamics as a descent in the loss landscape, drawing parallels with granular matter jamming. Analyze theoretical phase diagrams and empirical tests using MNIST parity. Investigate landscape curvature, flat directions, and the potential for overfitting. Consider ensemble averages and how fluctuations impact error rates. Discover scaling arguments and explore infinitely-wide networks, including initialization and learning processes. Gain insights into the Neural Tangent Kernel and finite N asymptotics. Enhance your understanding of deep learning theory and its connections to statistical physics.

Loss Landscape and Performance in Deep Learning by Stefano Spigler

International Centre for Theoretical Sciences
Add to list