Главная
Study mode:
on
1
- Intro & Overview
2
- Short Recap of Reinforcement Learning
3
- Problems with Model-Free Reinforcement Learning
4
- How World Models Help
5
- World Model Learner Architecture
6
- Deterministic & Stochastic Hidden States
7
- Latent Categorical Variables
8
- Categorical Variables and Multi-Modality
9
- Sampling & Stochastic State Prediction
10
- Actor-Critic Learning in Dream Space
11
- The Incompleteness of Learned World Models
12
- How General is this Algorithm?
13
- World Model Loss Function
14
- KL Balancing
15
- Actor-Critic Loss Function
16
- Straight-Through Estimators for Sampling Backpropagation
17
- Experimental Results
18
- Where Does It Fail?
19
- Conclusion
Description:
Explore a comprehensive video lecture on the Dreamer v2 algorithm, a groundbreaking approach in model-based reinforcement learning for mastering Atari games. Delve into the intricacies of world models, discrete representations, and latent space predictions as the speaker breaks down this collaborative research from Google AI, DeepMind, and the University of Toronto. Learn about the innovative use of discrete and stochastic latent states, the architecture of the world model learner, and the application of actor-critic learning in dream space. Gain insights into advanced concepts such as KL balancing, straight-through estimators, and the challenges of incomplete world models. Understand the experimental results, limitations, and potential applications of this state-of-the-art algorithm that achieves human-level performance on the Atari benchmark using a single GPU.

Mastering Atari with Discrete World Models - Machine Learning Research Paper Explained

Yannic Kilcher
Add to list
0:00 / 0:00