Главная
Study mode:
on
1
Intro
2
What is offline reinforcement learning
3
Intuition
4
Realizability
5
Sequential Decision Making
6
Standard Approach
7
Coverage
8
Limits
9
Policy Evaluation
10
Setting
11
Feature Mapping
12
Upper Limits
13
Lower Limits
14
Observations
15
Upper Bounds
16
Inequality
17
Simulation
18
Summary
19
Sufficient Conditions
20
Possible Results
21
Intuition and Construction
22
Practical Considerations
23
Follow Up
24
Experiments
25
Other Experiments
26
Model vs Feature
Description:
Explore the statistical boundaries of offline reinforcement learning with function approximation in this 55-minute lecture by Sham Kakade from the University of Washington and Microsoft Research. Delve into key concepts including realizability, sequential decision making, coverage limits, and policy evaluation. Examine upper and lower bounds, practical considerations, and experimental results. Gain insights into the mathematics of online decision making and the interplay between models and features in reinforcement learning.

What Are the Statistical Limits of Offline Reinforcement Learning With Function Approximation?

Simons Institute
Add to list
0:00 / 0:00