Главная
Study mode:
on
1
Intro
2
Main motivating examples
3
Subgradient method/descent (Shor et al., 1985)
4
Subgradient descent for machine learning Assumptions is the expected risk, the empirical risk
5
Summary: minimizing convex functions
6
Relationship to online learning
7
Stochastic subgradient "descent" /method
8
Convex stochastic approximation Existing work • Known global minimax rates of convergence for non-smooth problems (Nemirovsky and Yudin, 1983; Agarwal et al., 2012)
9
Robustness to wrong constants for = Cn
10
Robustness to lack of strong convexity
11
Beyond stochastic gradient method
12
Outline
13
Adaptive algorithm for logistic regression
14
Self-concordance
15
Least-mean-square algorithm
16
Markov chain interpretation of constant step sizes
17
Least-squares - Proof technique
18
Simulations - synthetic examples
Description:
Explore the intricacies of large-scale machine learning and convex optimization in this comprehensive lecture. Delve into the challenges of handling big data in machine learning and signal processing, focusing on online algorithms like stochastic gradient descent. Examine the optimal convergence rates for general convex and strongly-convex functions. Discover how the smoothness of loss functions can be leveraged to design innovative algorithms with improved performance. Learn about a novel Newton-based stochastic approximation algorithm that achieves faster convergence rates without strong convexity assumptions. Investigate the practical applications of combining batch and online algorithms for strongly convex problems. Cover topics such as subgradient descent, stochastic approximation, adaptive algorithms for logistic regression, self-concordance, and least-mean-square algorithms. Gain insights through theoretical proofs and synthetic example simulations.

Large Scale Machine Learning and Convex Optimization - Lecture 3

Hausdorff Center for Mathematics
Add to list
0:00 / 0:00