SOTAVerified

Option Discovery using Deep Skill Chaining

2020-05-01ICLR 2020Code Available1· sign in to hype

Akhil Bagaria, George Konidaris

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Autonomously discovering temporally extended actions, or skills, is a longstanding goal of hierarchical reinforcement learning. We propose a new algorithm that combines skill chaining with deep neural networks to autonomously discover skills in high-dimensional, continuous domains. The resulting algorithm, deep skill chaining, constructs skills with the property that executing one enables the agent to execute another. We demonstrate that deep skill chaining significantly outperforms both non-hierarchical agents and other state-of-the-art skill discovery techniques in challenging continuous control tasks.

Tasks

Reproductions