SOTAVerified

Globally Optimal Hierarchical Reinforcement Learning for Linearly-Solvable Markov Decision Processes

2021-06-29Code Available0· sign in to hype

Guillermo Infante, Anders Jonsson, Vicenç Gómez

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

In this work we present a novel approach to hierarchical reinforcement learning for linearly-solvable Markov decision processes. Our approach assumes that the state space is partitioned, and the subtasks consist in moving between the partitions. We represent value functions on several levels of abstraction, and use the compositionality of subtasks to estimate the optimal values of the states in each partition. The policy is implicitly defined on these optimal value estimates, rather than being decomposed among the subtasks. As a consequence, our approach can learn the globally optimal policy, and does not suffer from the non-stationarity of high-level decisions. If several partitions have equivalent dynamics, the subtasks of those partitions can be shared. If the set of boundary states is smaller than the entire state space, our approach can have significantly smaller sample complexity than that of a flat learner, and we validate this empirically in several experiments.

Tasks

Reproductions