Single-hidden layer (shallow) neural networks of infinite width Consider a NN which
5
What GP does it correspond to?
6
Properties of the NNGP
7
Bayesian inference with a GP prior (Review)
8
Experiments from original work
9
Performance comparison
10
NNGP performance across hyperparameters
11
Large depth behavior & fixed points
12
Phase diagrams: experiments vs. theory
13
Performance trends with width and dataset size
14
Empirical comparison of various NN-GPS
15
Empirical trends
16
Best performing networks: comparison between GPs and SGD-NNS
17
Partway summary
18
What dynamics occurs in parameter space?
19
Closing Remarks
Description:
Explore the computational aspects of extremely wide neural networks in this 49-minute lecture by Yasaman Bahri from Google Brain. Delve into topics such as single-hidden layer neural networks of infinite width, Gaussian Process (GP) correspondences, and Bayesian inference with GP priors. Examine experimental results comparing neural network Gaussian process (NNGP) performance across hyperparameters, large depth behavior, and fixed points. Analyze phase diagrams, performance trends with width and dataset size, and empirical comparisons of various NN-GPs. Investigate the dynamics occurring in parameter space and gain insights into the best-performing networks, comparing GPs and SGD-trained neural networks. Part of the Frontiers of Deep Learning series at the Simons Institute.