Backpropagation through the Void: Optimizing control variates for black-box gradient estimation
Will Grathwohl, Dami Choi, Yuhuai Wu, Geoffrey Roeder, David Duvenaud
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/duvenaud/relaxOfficialIn papertf★ 0
- github.com/thlautenschlaeger/bpttv-laxpytorch★ 1
- github.com/brain-research/mirage-rltf★ 0
- github.com/wgrathwohl/BackpropThroughTheVoidRLtf★ 0
- github.com/ElleryL/gradient_estimatorpytorch★ 0
- github.com/Bonnevie/rebartf★ 0
- github.com/TalkToTheGAN/REGANpytorch★ 0
Abstract
Gradient-based optimization is the foundation of deep learning and reinforcement learning. Even when the mechanism being optimized is unknown or not differentiable, optimization using high-variance or biased gradient estimates is still often the best strategy. We introduce a general framework for learning low-variance, unbiased gradient estimators for black-box functions of random variables. Our method uses gradients of a neural network trained jointly with model parameters or policies, and is applicable in both discrete and continuous settings. We demonstrate this framework for training discrete latent-variable models. We also give an unbiased, action-conditional extension of the advantage actor-critic reinforcement learning algorithm.