Главная
Study mode:
on
1
Intro
2
Statistical models come to rescue
3
Example: low-rank matrix recovery
4
Solving quadratic systems of equations
5
A natural least squares formulation
6
Rationale of two-stage approach
7
What does prior theory say?
8
Exponential growth of signal strength in Stage 1
9
Our theory: noiseless case
10
Population-level state evolution
11
Back to finite-sample analysis
12
Gradient descent theory revisited
13
A second look at gradient descent theory
14
Key proof idea: leave-one-out analysis
15
Key proof ingredient: random-sign sequences
16
Automatic saddle avoidance
Description:
Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only! Grab it Explore the second lecture in a five-part series featuring Princeton University's Yuxin Chen, focusing on random initialization and implicit regularization in nonconvex statistical estimation. Delve into the phenomenon where gradient descent converges to optimal solutions in nonconvex problems like phase retrieval and matrix completion, achieving near-optimal statistical and computational guarantees without careful initialization or explicit regularization. Examine the leave-one-out approach used to decouple statistical dependency between gradient descent iterates and data. Learn about the application of this method to noisy matrix completion, demonstrating near-optimal entrywise error control. Investigate topics such as low-rank matrix recovery, quadratic systems of equations, two-stage approaches, population-level state evolution, and automatic saddle avoidance in this 48-minute talk from the TRIAD Distinguished Lecture Series at Georgia Tech Research.

Random Initialization and Implicit Regularization in Nonconvex Statistical Estimation - Lecture 2

Georgia Tech Research
Add to list
0:00 / 0:00