SOTAVerified

A Finite Sample Complexity Bound for Distributionally Robust Q-learning

2023-02-26Unverified0· sign in to hype

Shengbo Wang, Nian Si, Jose Blanchet, Zhengyuan Zhou

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

We consider a reinforcement learning setting in which the deployment environment is different from the training environment. Applying a robust Markov decision processes formulation, we extend the distributionally robust Q-learning framework studied in Liu et al. [2022]. Further, we improve the design and analysis of their multi-level Monte Carlo estimator. Assuming access to a simulator, we prove that the worst-case expected sample complexity of our algorithm to learn the optimal robust Q-function within an error in the sup norm is upper bounded by O(|S||A|(1-)^-5^-2p_^-6^-4), where is the discount rate, p_ is the non-zero minimal support probability of the transition kernels and is the uncertainty size. This is the first sample complexity result for the model-free robust RL problem. Simulation studies further validate our theoretical results.

Tasks

Reproductions