Control-Oriented Model-Based Reinforcement Learning with Implicit Differentiation
9
Optimal Model Design Problem (OMD)
10
Smooth Bellman Optimality Equations
11
Connection between OMD and Rust (1988)
12
Bilevel Optimization (Bard 1998)
13
Implicit and Iterative Differentiation
14
Benefits under Model Misspecification
15
Function Approximation and Distractor States
16
Performance under Model Misspecification
17
Continuous-Time Meta-Learning with Forward Mode Dif- ferentiation.
18
Gradient Flow-based Meta-Learning
19
Time irreversibility
20
Memory-efficient meta-gradients
21
Consequence
22
Empirical Efficiency of COML
23
Nonlinear Trajectory Optimization
24
Extragradient Method
25
Trajectory Optimization with Learned Model
Description:
Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Grab it
Explore decision awareness in reinforcement learning through this 54-minute seminar presented by Pierre-Luc Bacon from Université de Montréal. Delve into the end-to-end perspective of optimizing learning systems for optimal decision-making, focusing on recent advances in model-based reinforcement learning. Examine control-oriented transition models using implicit differentiation and the application of neural ordinary differential equations for nonlinear trajectory optimization. Investigate computational challenges and scaling solutions, including efficient Jacobian factorization in forward mode automatic differentiation and novel constrained optimizers inspired by adversarial learning. Cover topics such as optimal model design, smooth Bellman optimality equations, bilevel optimization, continuous-time meta-learning, and gradient flow-based techniques.
Decision Awareness in Reinforcement Learning - End-to-End Optimization Approaches