SOTAVerified

Rényi Divergence Variational Inference

2016-02-06NeurIPS 2016Code Available0· sign in to hype

Yingzhen Li, Richard E. Turner

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

This paper introduces the variational R\'enyi bound (VR) that extends traditional variational inference to R\'enyi's alpha-divergences. This new family of variational methods unifies a number of existing approaches, and enables a smooth interpolation from the evidence lower-bound to the log (marginal) likelihood that is controlled by the value of alpha that parametrises the divergence. The reparameterization trick, Monte Carlo approximation and stochastic optimisation methods are deployed to obtain a tractable and unified framework for optimisation. We further consider negative alpha values and propose a novel variational inference method as a new special case in the proposed framework. Experiments on Bayesian neural networks and variational auto-encoders demonstrate the wide applicability of the VR bound.

Tasks

Reproductions