SOTAVerified

Sequential Function-Space Variational Inference via Gaussian Mixture Approximation

2025-03-10Code Available0· sign in to hype

Menghao Waiyan William Zhu, Pengcheng Hao, Ercan Engin Kuruoğlu

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Continual learning is learning from a sequence of tasks with the aim of learning new tasks without forgetting old tasks. Sequential function-space variational inference (SFSVI) is a continual learning method based on variational inference which uses a Gaussian variational distribution to approximate the distribution of the outputs of a finite number of selected inducing points. Since the posterior distribution of a neural network is multi-modal, a Gaussian distribution could only match one mode of the posterior distribution, and a Gaussian mixture distribution could be used to better approximate the posterior distribution. We propose an SFSVI method which uses a Gaussian mixture variational distribution. We also compare different types of variational inference methods with and without a fixed pre-trained feature extractor. We find that in terms of final average accuracy, Gaussian mixture methods perform better than Gaussian methods and likelihood-focused methods perform better than prior-focused methods.

Tasks

Reproductions