SOTAVerified

Minimum Description Length Skills for Accelerated Reinforcement Learning

2021-03-09ICLR Workshop SSL-RL 2021Unverified0· sign in to hype

Jesse Zhang, Karl Pertsch, Jiefan Yang, Joseph J Lim

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

Humans can quickly learn new tasks by reusing a large number of previously acquired skills. How can we discover such reusable skills for artificial agents when given a large dataset of prior experience? Past works leverage extensive human supervision to define skills or use simple skill heuristics that limit their expressiveness. In contrast, we propose a principled, unsupervised objective for skill discovery from large, offline datasets based on the Minimum Description Length principle: we show that a "code book" of skills that can maximally compress the training data can be reused to efficiently learn new tasks. By minimizing description length we strike an optimal balance between the number of extracted skills and their complexity. We show that our approach outperforms alternative approaches that heuristically define skills on a complex, long-horizon maze navigation task.

Tasks

Reproductions