SOTAVerified

Learning Good Policies By Learning Good Perceptual Models

2019-09-25Unverified0· sign in to hype

Yilun Du, Phillip Isola

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

Reinforcement learning (RL) has led to increasingly complex looking behavior in recent years. However, such complexity can be misleading and hides over-fitting. We find that visual representations may be a useful metric of complexity, and both correlates well objective optimization and causally effects reward optimization. We then propose curious representation learning (CRL) which allows us to use better visual representation learning algorithms to correspondingly increase visual representation in policy through an intrinsic objective on both simulated environments and transfer to real images. Finally, we show better visual representations induced by CRL allows us to obtain better performance on Atari without any reward than other curiosity objectives.

Tasks

Reproductions