Главная
Study mode:
on
1
Intro
2
Deep learning helps us handle unstructured environments
3
Reinforcement learning provides a formalism for behavior
4
RL has a big problem
5
Off-policy RL with large datasets
6
Off-policy model-free learning
7
How to solve for the Q-function?
8
QT-Opt: off-policy Q-learning at scale
9
Grasping with QT-Opt
10
Emergent grasping strategies
11
So what's the problem?
12
How to stop training on garbage?
13
How well does it work?
14
Off-policy model-based reinforcement learning
15
High-level algorithm outline
16
Model-based RL for dexterous manipulation
17
Q-Functions (can) learn models
18
Temporal difference models
19
Optimizing over valid states
Description:
Explore deep reinforcement learning applications in real-world scenarios through this insightful lecture by Sergey Levine from the University of Berkeley. Delve into the challenges and solutions of off-policy reinforcement learning with large datasets, focusing on model-free and model-based approaches. Learn about QT-Opt, an off-policy Q-learning algorithm at scale, and its application in robotic grasping tasks. Discover how to address common issues in reinforcement learning, such as training on irrelevant data, and understand the potential of temporal difference models and Q-functions in learning implicit models. Gain valuable insights into optimizing over valid states and the application of model-based reinforcement learning for dexterous manipulation tasks.

Deep Reinforcement Learning in the Real World - Sergey Levine

Institute for Advanced Study
Add to list
0:00 / 0:00