SOTAVerified

UVIP: Model-Free Approach to Evaluate Reinforcement Learning Algorithms

2021-05-05Code Available0· sign in to hype

Ilya Levin, Denis Belomestny, Alexey Naumov, Sergey Samsonov

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Policy evaluation is an important instrument for the comparison of different algorithms in Reinforcement Learning (RL). Yet even a precise knowledge of the value function V^ corresponding to a policy does not provide reliable information on how far is the policy from the optimal one. We present a novel model-free upper value iteration procedure ( UVIP) that allows us to estimate the suboptimality gap V^(x) - V^(x) from above and to construct confidence intervals for V^. Our approach relies on upper bounds to the solution of the Bellman optimality equation via martingale approach. We provide theoretical guarantees for UVIP under general assumptions and illustrate its performance on a number of benchmark RL problems.

Tasks

Reproductions