Главная
Study mode:
on
1
Intro
2
Empirical RL for large-scale problems
3
Traditional Online RL paradigm
4
offline RL paradigm
5
Potential Applications of Offline RL
6
Why offline RL is challenging?
7
Finite Horizon MDPs
8
Offline Data Collection
9
Offline Data Coverage
10
Learning goal in Offline RL: Robustness
11
Learning goal in Offline RL: Generalization
12
A Model-based Approach
13
A Naive Model-based Approach
14
Formal Theoretical Guarantee for CPPO
15
Traditional Imitation Learning
16
Covariate shift in Imitation
17
Offline Imitation Learning
18
Explanation of MILO
19
Experiments of MILO
Description:
Explore a comprehensive lecture on offline reinforcement learning, focusing on generalization and robustness. Delve into the challenges and potential applications of this learning paradigm, which uses pre-collected static datasets without further environment interaction. Examine a general model-based offline RL algorithm that demonstrates generalization in large-scale Markov Decision Processes and robustness in policy discovery. Investigate offline Imitation Learning, including an algorithm with polynomial sample complexity and state-of-the-art performance in continuous control robotics benchmarks. Cover topics such as empirical RL for large-scale problems, finite horizon MDPs, offline data collection and coverage, learning goals in offline RL, and traditional versus offline imitation learning.

Generalization and Robustness in Offline Reinforcement Learning

Simons Institute
Add to list