Conference Program Home
  My Program

All Times EDT

Abstract Details

Activity Number: 309 - Statistical Reinforcement Learning
Type: Invited
Date/Time: Tuesday, August 9, 2022 : 2:00 PM to 3:50 PM
Sponsor: Section on Statistical Computing
Abstract #320353
Title: Settling the Sample Complexity of Model-Based Offline Reinforcement Learning
Author(s): Yuxin Chen* and Yuxin Chen and Laixi Shi and Yuejie Chi and Yuting Wei
Companies: Princeton University and University of Pennsylvania and Carnegie Mellon University and Carnegie Mellon University and University of Pennsylvania
Keywords: model-based RL; reinforcement learning; sample complexity; pessimism
Abstract:

This work is concerned with offline reinforcement learning (RL), which learns using pre-collected data without further exploration. Effective offline RL would be able to accommodate distribution shift and insufficient data coverage. However, prior algorithms either suffer from suboptimal sample complexities or incur high burn-in cost, thus posing an impediment to efficient RL in sample-starved applications. In this work, we demonstrate that the model-based (or ``plug-in'') approach achieves minimax-optimal sample optimality with minimal burn-in cost. Our algorithms are pessimistic variants of value iteration with Bernstein-style penalties, which do not rely on sophisticated schemes like variance reduction.


Authors who are presenting talks have a * after their name.

Back to the full JSM 2022 program