SOTAVerified

Bayesian Policy Gradients via Alpha Divergence Dropout Inference

2017-12-06Code Available0· sign in to hype

Peter Henderson, Thang Doan, Riashat Islam, David Meger

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Policy gradient methods have had great success in solving continuous control tasks, yet the stochastic nature of such problems makes deterministic value estimation difficult. We propose an approach which instead estimates a distribution by fitting the value function with a Bayesian Neural Network. We optimize an -divergence objective with Bayesian dropout approximation to learn and estimate this distribution. We show that using the Monte Carlo posterior mean of the Bayesian value function distribution, rather than a deterministic network, improves stability and performance of policy gradient methods in continuous control MuJoCo simulations.

Tasks

Reproductions