Главная
Study mode:
on
1
- Listing the neural networks we need to train
2
- What a training batch item looks like
3
- Forward passes and losses
4
- Why the latent state representation does not collapse
5
- Understanding TD Learning
6
- TD learning intuition in real experiments
7
- Optimizing the Q network using the TD error
8
- Offline vs online data collection and training loop
9
- Wrapping up
Description:
Learn about training neural networks in TD-MPC (Temporal Difference Model Predictive Control) through a 38-minute technical video lecture that delves into the implementation details and theoretical foundations. Explore the training process by examining neural network requirements, batch structure, forward passes, and loss calculations. Understand key concepts like latent state representation stability, TD Learning principles, and their practical applications in real experiments. Master the optimization of Q networks using TD error and differentiate between offline and online data collection approaches. Access referenced research papers and implementation code from the LeRobot library while benefiting from detailed explanations across multiple topics, from basic network training to advanced concepts in temporal difference learning.

Training Neural Networks for Temporal Difference Model Predictive Control - Part 2

HuggingFace
Add to list
0:00 / 0:00