Главная
Study mode:
on
1
Intro
2
Tensor (CP) decomposition
3
Why naïve algorithm fails
4
Why gradient descent?
5
Two-Layer Neural Network
6
Form of the objective
7
Difficulties of analyzing gradient descent
8
Lazy training fails
9
O is a high order saddle point
10
Our (high level) algorithm
11
Proof ideas
12
Iterates remain close to correct subspace
13
Escaping local minima by random correlation
14
Amplify initial correlation by tensor power method
15
Conclusions and Open Problems
Description:
Explore a lecture on over-parameterized tensor decomposition and its applications beyond lazy training. Delve into the mathematical foundations and algorithms for tensor computations, focusing on how gradient descent variants can find approximate tensor decompositions. Learn about the limitations of lazy training regimes, the challenges in analyzing gradient descent, and a novel high-level algorithm that overcomes these obstacles. Discover how this research relates to training neural networks and utilizing low-rank structure in data. Gain insights into the proof ideas, including maintaining iterates close to the correct subspace and escaping local minima through random correlation and tensor power methods.

Beyond Lazy Training for Over-parameterized Tensor Decomposition

Institute for Pure & Applied Mathematics (IPAM)
Add to list
0:00 / 0:00