SOTAVerified

Task Agnostic Continual Learning Using Online Variational Bayes

2018-03-27Code Available0· sign in to hype

Chen Zeno, Itay Golan, Elad Hoffer, Daniel Soudry

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Catastrophic forgetting is the notorious vulnerability of neural networks to the change of the data distribution while learning. This phenomenon has long been considered a major obstacle for allowing the use of learning agents in realistic continual learning settings. A large body of continual learning research assumes that task boundaries are known during training. However, research for scenarios in which task boundaries are unknown during training has been lacking. In this paper we present, for the first time, a method for preventing catastrophic forgetting (BGD) for scenarios with task boundaries that are unknown during training --- task-agnostic continual learning. Code of our algorithm is available at https://github.com/igolan/bgd.

Tasks

Reproductions