Kalman Gradient Descent: Adaptive Variance Reduction in Stochastic Optimization
2018-10-29Code Available0· sign in to hype
James Vuckovic
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/jamesvuc/KGDOfficialIn papertf★ 0
Abstract
We introduce Kalman Gradient Descent, a stochastic optimization algorithm that uses Kalman filtering to adaptively reduce gradient variance in stochastic gradient descent by filtering the gradient estimates. We present both a theoretical analysis of convergence in a non-convex setting and experimental results which demonstrate improved performance on a variety of machine learning areas including neural networks and black box variational inference. We also present a distributed version of our algorithm that enables large-dimensional optimization, and we extend our algorithm to SGD with momentum and RMSProp.