Главная
Study mode:
on
1
Multiarmed bandits
2
Exploration exploitation
3
Stochastic bandits
4
Bandits from gambling
5
Bandits in practice
6
Online optimization
7
Simplified version
8
The problem
9
Heuristics
10
Notion of regret
11
Epsilon greedy strategy
12
Single state
13
Epsilon greedy
14
Different approaches
15
In practice
Description:
Explore the fascinating world of multi-armed bandits in this comprehensive 57-minute lecture by Pascal Poupart. Delve into key concepts such as exploration-exploitation trade-offs, stochastic bandits, and online optimization. Learn about the origins of bandits in gambling and their practical applications. Understand the simplified version of the problem, various heuristics, and the notion of regret. Discover the epsilon-greedy strategy and its implementation in single-state scenarios. Gain insights into different approaches and their effectiveness in real-world situations.

CS885: Multi-Armed Bandits

Pascal Poupart
Add to list
0:00 / 0:00