Adversarial Online Multi-Task Reinforcement Learning
Quan Nguyen, Nishant A. Mehta
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/ngmq/adversarial-online-multi-task-reinforcement-learningOfficialIn papernone★ 3
Abstract
We consider the adversarial online multi-task reinforcement learning setting, where in each of K episodes the learner is given an unknown task taken from a finite set of M unknown finite-horizon MDP models. The learner's objective is to minimize its regret with respect to the optimal policy for each task. We assume the MDPs in M are well-separated under a notion of -separability, and show that this notion generalizes many task-separability notions from previous works. We prove a minimax lower bound of (KDSAH) on the regret of any learning algorithm and an instance-specific lower bound of (K^2) in sample complexity for a class of uniformly-good cluster-then-learn algorithms. We use a novel construction called 2-JAO MDP for proving the instance-specific lower bound. The lower bounds are complemented with a polynomial time algorithm that obtains O(K^2) sample complexity guarantee for the clustering phase and O(MK) regret guarantee for the learning phase, indicating that the dependency on K and 1^2 is tight.