Главная
Study mode:
on
1
Introduction
2
Parametric Statistics
3
Maximum likelihood estimation
4
How to constrain noise
5
Estimating parameters
6
Do empirical mean and empirical variance work
7
Folklore Theorem
8
Robustness and Hardness
9
Price of Robustness
10
Recent Results
11
Robust Estimation Recipe
12
WinWin Algorithm
13
Birds Eye View
14
O of epsilon
15
Relaxing Distributional Assumption
16
Robust Estimation
17
Conclusion
18
Does the error guarantee tend to O
19
Is there a sense that some adversaries are complete
20
gaussian mean estimation
21
adversary efficiency
22
improper learning
Description:
Explore recent advancements in high-dimensional learning through this comprehensive lecture by MIT's Ankur Moitra. Delve into parametric statistics, maximum likelihood estimation, and noise constraint techniques. Examine the effectiveness of empirical mean and variance, and understand the Folklore Theorem. Investigate robustness, hardness, and the price of robustness in estimation. Learn about the Robust Estimation Recipe and the WinWin Algorithm. Analyze relaxed distributional assumptions and their impact on robust estimation. Consider the error guarantee tendencies and the completeness of adversaries in gaussian mean estimation. Gain insights into adversary efficiency and improper learning in high-dimensional contexts.

Recent Progress in High-Dimensional Learning

Simons Institute
Add to list
0:00 / 0:00