SOTAVerified

Goal-Driven Imitation Learning from Observation by Inferring Goal Proximity

2021-01-01Unverified0· sign in to hype

Andrew Szot, Youngwoon Lee, Shao-Hua Sun, Joseph J Lim

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

Humans can effectively learn to estimate how close they are to completing a desired task simply by watching others fulfill the task. To solve the task, they can then take actions towards states with higher estimated proximity to the goal. From this intuition, we propose a simple yet effective method for imitation learning that learns a goal proximity function from expert demonstrations and online agent experience, and then uses the learned proximity to provide a dense reward signal for training a policy to solve the task. By predicting task progress as the temporal distance to the goal, the goal proximity function improves generalization to unseen states over methods that aim to directly imitate expert behaviors. We demonstrate that our proposed method efficiently learns a set of goal-driven tasks from state-only demonstrations in navigation, robotic arm manipulation, and locomotion tasks.

Tasks

Reproductions