SOTAVerified

Visual Transfer for Reinforcement Learning via Wasserstein Domain Confusion

2020-06-04Code Available0· sign in to hype

Josh Roy, George Konidaris

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

We introduce Wasserstein Adversarial Proximal Policy Optimization (WAPPO), a novel algorithm for visual transfer in Reinforcement Learning that explicitly learns to align the distributions of extracted features between a source and target task. WAPPO approximates and minimizes the Wasserstein-1 distance between the distributions of features from source and target domains via a novel Wasserstein Confusion objective. WAPPO outperforms the prior state-of-the-art in visual transfer and successfully transfers policies across Visual Cartpole and two instantiations of 16 OpenAI Procgen environments.

Tasks

Reproductions