Unbiased Policy Evaluation for General RL in Short Horizons
5
Queue-based Offline Evaluation of Online Bandit Algorithms
6
Our Queue Approach Can Sometimes Evaluate Algorithms that Use Deterministic Policies for Many More Time Steps than Rejection
7
Sample Complexity of RL
8
Provably More Efficient Learners
9
Fast, Better Policy Search using Bayesian Optimization
10
Black Box Optimization
11
Opening the Box: Leverage Offline Policy Evaluation
12
Personalization & Transfer Learning for Sequential Decision Making Tasks
13
Latent Variable Modeling Background
14
Diameter Assumption: Needed for Sample Complexity Improvement in Transfer?
15
Active Set is Models Compatible with Current Task's Data
16
More Data Efficient Learning In Domains Where It Matters
Description:
Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Grab it
Explore cutting-edge research on creating fast autonomous learners in this 57-minute lecture by Emma Brunskill at the Center for Language & Speech Processing, JHU. Delve into the challenges of developing AI agents that can make good decisions in stochastic environments, with a focus on applications involving human interaction. Learn about transfer learning across sequential decision-making tasks and its potential to improve educational tools. Discover key concepts such as Markov Decision Processes, reinforcement learning, and unbiased policy evaluation. Examine innovative approaches like queue-based offline evaluation and Bayesian optimization for faster, more efficient policy search. Investigate personalization and transfer learning techniques for sequential decision-making tasks, and understand the importance of sample efficiency in real-world applications. Gain insights from Brunskill's expertise in AI and machine learning, particularly in the context of intelligent tutoring systems and medical interfaces.
Read more
Towards Fast Autonomous Learners: Advances in Reinforcement Learning - 2015
Center for Language & Speech Processing(CLSP), JHU