SOTAVerified

A Latent Variational Framework for Stochastic Optimization

2019-05-05NeurIPS 2019Unverified0· sign in to hype

Philippe Casgrain

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

This paper provides a unifying theoretical framework for stochastic optimization algorithms by means of a latent stochastic variational problem. Using techniques from stochastic control, the solution to the variational problem is shown to be equivalent to that of a Forward Backward Stochastic Differential Equation (FBSDE). By solving these equations, we recover a variety of existing adaptive stochastic gradient descent methods. This framework establishes a direct connection between stochastic optimization algorithms and a secondary Bayesian inference problem on gradients, where a prior measure on noisy gradient observations determines the resulting algorithm.

Tasks

Reproductions