Online Regret Bounds for Undiscounted Continuous Reinforcement Learning
2013-02-11NeurIPS 2012Unverified0· sign in to hype
Ronald Ortner, Daniil Ryabko
Unverified — Be the first to reproduce this paper.
ReproduceAbstract
We derive sublinear regret bounds for undiscounted reinforcement learning in continuous state space. The proposed algorithm combines state aggregation with the use of upper confidence bounds for implementing optimism in the face of uncertainty. Beside the existence of an optimal policy which satisfies the Poisson equation, the only assumptions made are Holder continuity of rewards and transition probabilities.