SOTAVerified

Variational Inference for Policy Gradient

2018-02-21Unverified0· sign in to hype

Tianbing Xu

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

Inspired by the seminal work on Stein Variational Inference and Stein Variational Policy Gradient, we derived a method to generate samples from the posterior variational parameter distribution by explicitly minimizing the KL divergence to match the target distribution in an amortize fashion. Consequently, we applied this varational inference technique into vanilla policy gradient, TRPO and PPO with Bayesian Neural Network parameterizations for reinforcement learning problems.

Tasks

Reproductions