What might generalization theory offer deep learning?
3
Barriers to explaining generalization
4
PAC-Bayes yields risk bounds for Gibbs classifiers
5
PAC-Bayes generalization bounds
6
PAC-Bayes bounds on deterministic classifiers
7
Distribution-dependent approximations of optimal priors via privacy
8
A question of interpretation
9
Use SGD to predict SGD
10
Data and distribution priors for neural networks
11
MNIST Results - Coupled data dependent priors and posteriors
12
Oracle access to optimal prior covariance
13
Bounds with oracle covariance + ghost sample
14
Bounds on 32k samples v 64k samples
15
Recap and Conclusion
Description:
Explore the intersection of generalization theory and deep learning in this 45-minute lecture from the Frontiers of Deep Learning series. Delve into PAC-Bayes theory and its applications to risk bounds for Gibbs classifiers and deterministic classifiers. Examine distribution-dependent approximations of optimal priors, the role of privacy, and the use of SGD to predict SGD. Investigate data and distribution priors for neural networks, focusing on MNIST results with coupled data-dependent priors and posteriors. Analyze bounds with oracle covariance and ghost samples, comparing results across different sample sizes. Gain insights into the potential contributions of generalization theory to deep learning and the challenges in explaining generalization.
Studying Generalization in Deep Learning via PAC-Bayes