Gaussian Processes for Big Data
2013-09-26Code Available2· sign in to hype
James Hensman, Nicolo Fusi, Neil D. Lawrence
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/GPflow/GPflowtf★ 1,906
- github.com/ratschlab/SVGP-VAEtf★ 22
- github.com/danehuang/softki_gp_kitpytorch★ 2
- github.com/SheffieldML/GPyIn papernone★ 0
- github.com/secondmind-labs/bayesian_benchmarksnone★ 0
- github.com/cornellius-gp/gpytorchpytorch★ 0
- github.com/hughsalimbeni/bayesian_benchmarksnone★ 0
- github.com/UnofficialJuliaMirrorSnapshots/AugmentedGaussianProcesses.jl-38eea1fd-7d7d-5162-9d08-f89d0f2e271enone★ 0
- github.com/UnofficialJuliaMirror/AugmentedGaussianProcesses.jl-38eea1fd-7d7d-5162-9d08-f89d0f2e271enone★ 0
Abstract
We introduce stochastic variational inference for Gaussian process models. This enables the application of Gaussian process (GP) models to data sets containing millions of data points. We show how GPs can be vari- ationally decomposed to depend on a set of globally relevant inducing variables which factorize the model in the necessary manner to perform variational inference. Our ap- proach is readily extended to models with non-Gaussian likelihoods and latent variable models based around Gaussian processes. We demonstrate the approach on a simple toy problem and two real world data sets.