SOTAVerified

Distributional Reinforcement Learning with Monotonic Splines

2021-09-29ICLR 2022Unverified0· sign in to hype

Yudong Luo, Guiliang Liu, Haonan Duan, Oliver Schulte, Pascal Poupart

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

Distributional Reinforcement Learning (RL) differs from traditional RL by estimating the distribution over returns to capture the intrinsic uncertainty of MDPs. One key challenge in distributional RL lies in how to parameterize the quantile function when minimizing the Wasserstein metric of temporal differences. Existing algorithms use a step functions or piecewise linear functions. In this paper, we propose to learn smooth continuous quantile functions represented by monotonic rational-quadratic splines, which also naturally solve the quantile crossing problem. Experiments in stochastic environments show that a dense estimation for quantile functions enhances distributional RL in terms of faster empirical convergence and higher rewards in most cases.

Tasks

Reproductions