Главная
Study mode:
on
1
Core" of the data set • Observe: redundancies can exist even if data isn't "tall
2
Roadmap
3
Bayesian coresets
4
Uniform subsampling revisited
5
Importance sampling
6
Hilbert coresets
7
Frank-Wolfe
8
Gaussian model (simulated) • 1K pts; norms, inference: closed-form
9
Logistic regression (simulated)
10
Real data experiments
11
Data summarization
Description:
Explore advanced techniques for scalable Bayesian inference in large-scale data settings through this 59-minute lecture by MIT's Tamara Broderick. Delve into the concept of data summarization and coresets as a means to overcome computational challenges in Bayesian methods. Learn about theoretical guarantees on coreset size and approximation quality, and discover how this approach provides geometric decay in posterior approximation error. Examine the application of these techniques to both synthetic and real datasets, demonstrating significant improvements over uniform random subsampling. Gain insights into Broderick's research on developing and analyzing models for scalable Bayesian machine learning, and understand the potential impact of these methods on handling large datasets efficiently while maintaining the benefits of Bayesian inference.

Automated Scalable Bayesian Inference via Data Summarization - 2018

Center for Language & Speech Processing(CLSP), JHU
Add to list
0:00 / 0:00