SOTAVerified

Model-based Reinforcement Learning and the Eluder Dimension

2014-06-07NeurIPS 2014Unverified0· sign in to hype

Ian Osband, Benjamin Van Roy

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

We consider the problem of learning to optimize an unknown Markov decision process (MDP). We show that, if the MDP can be parameterized within some known function class, we can obtain regret bounds that scale with the dimensionality, rather than cardinality, of the system. We characterize this dependence explicitly as O(d_K d_E T) where T is time elapsed, d_K is the Kolmogorov dimension and d_E is the eluder dimension. These represent the first unified regret bounds for model-based reinforcement learning and provide state of the art guarantees in several important settings. Moreover, we present a simple and computationally efficient algorithm posterior sampling for reinforcement learning (PSRL) that satisfies these bounds.

Tasks

Reproductions