Online Program Home
  My Program

All Times EDT

Abstract Details

Activity Number: 276 - Statistical Foundations of Reinforcement Learning
Type: Topic-Contributed
Date/Time: Wednesday, August 11, 2021 : 1:30 PM to 3:20 PM
Sponsor: IMS
Abstract #317062
Title: Learning Good State and Action Representations via Tensor Decomposition
Author(s): Anru Zhang* and Chengzhuo Ni and Yaqi Duan and Mengdi Wang
Companies: University of Wisconsin-Madison and Princeton University and Princeton University and Princeton University
Keywords: dimension reduction; low-Tucker-rank tensor; Markov decision process; unsupervised learning
Abstract:

The transition kernel of a continuous-state-action Markov decision process (MDP) admits a natural tensor structure. This paper proposes a tensor-inspired unsupervised learning method to identify meaningful low-dimensional state and action representations from empirical trajectories. The method exploits the MDP's tensor structure by kernelization, importance sampling and low-Tucker-rank approximation. This method can be further used to cluster states and actions respectively and find the best discrete MDP abstraction. We provide sharp statistical error bounds for tensor concentration and the preservation of diffusion distance after embedding. We further prove that the learned state/action abstractions provide accurate approximations to latent block structures if they exist, enabling function approximation in downstream tasks such as policy evaluation.


Authors who are presenting talks have a * after their name.

Back to the full JSM 2021 program