Visual Transfer for Reinforcement Learning via Wasserstein Domain Confusion
2020-06-04Code Available0· sign in to hype
Josh Roy, George Konidaris
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/ku2482/wappo.pytorchpytorch★ 7
Abstract
We introduce Wasserstein Adversarial Proximal Policy Optimization (WAPPO), a novel algorithm for visual transfer in Reinforcement Learning that explicitly learns to align the distributions of extracted features between a source and target task. WAPPO approximates and minimizes the Wasserstein-1 distance between the distributions of features from source and target domains via a novel Wasserstein Confusion objective. WAPPO outperforms the prior state-of-the-art in visual transfer and successfully transfers policies across Visual Cartpole and two instantiations of 16 OpenAI Procgen environments.