SOTAVerified

Distributional Reinforcement Learning with Regularized Wasserstein Loss

2022-02-01Code Available0· sign in to hype

Ke Sun, Yingnan Zhao, Wulong Liu, Bei Jiang, Linglong Kong

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

The empirical success of distributional reinforcement learning (RL) highly relies on the choice of distribution divergence equipped with an appropriate distribution representation. In this paper, we propose Sinkhorn distributional RL (SinkhornDRL), which leverages Sinkhorn divergence, a regularized Wasserstein loss, to minimize the difference between current and target Bellman return distributions. Theoretically, we prove the contraction properties of SinkhornDRL, aligning with the interpolation nature of Sinkhorn divergence between Wasserstein distance and Maximum Mean Discrepancy (MMD). The introduced SinkhornDRL enriches the family of distributional RL algorithms, contributing to interpreting the algorithm behaviors compared with existing approaches by our investigation into their relationships. Empirically, we show that SinkhornDRL consistently outperforms or matches existing algorithms on the Atari games suite and particularly stands out in the multi-dimensional reward setting. Code is available in https://github.com/datake/SinkhornDistRL..

Tasks

Reproductions