Online Linear Regression and Its Application to Model-Based Reinforcement Learning
2007-12-01NeurIPS 2007Unverified0· sign in to hype
Alexander L. Strehl, Michael L. Littman
Unverified — Be the first to reproduce this paper.
ReproduceAbstract
We provide a provably efficient algorithm for learning Markov Decision Processes (MDPs) with continuous state and action spaces in the online setting. Specifically, we take a model-based approach and show that a special type of online linear regression allows us to learn MDPs with (possibly kernalized) linearly parameterized dynamics. This result builds on Kearns and Singh's work that provides a provably efficient algorithm for finite state MDPs. Our approach is not restricted to the linear setting, and is applicable to other classes of continuous MDPs.