SOTAVerified

Towards Redundancy-Free Sub-networks in Continual Learning

2023-12-01Code Available0· sign in to hype

Cheng Chen, Jingkuan Song, Lianli Gao, Heng Tao Shen

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Catastrophic Forgetting (CF) is a prominent issue in continual learning. Parameter isolation addresses this challenge by masking a sub-network for each task to mitigate interference with old tasks. However, these sub-networks are constructed relying on weight magnitude, which does not necessarily correspond to the importance of weights, resulting in maintaining unimportant weights and constructing redundant sub-networks. To overcome this limitation, inspired by information bottleneck, which removes redundancy between adjacent network layers, we propose Information Bottleneck Masked sub-network (IBM) to eliminate redundancy within sub-networks. Specifically, IBM accumulates valuable information into essential weights to construct redundancy-free sub-networks, not only effectively mitigating CF by freezing the sub-networks but also facilitating new tasks training through the transfer of valuable knowledge. Additionally, IBM decomposes hidden representations to automate the construction process and make it flexible. Extensive experiments demonstrate that IBM consistently outperforms state-of-the-art methods. Notably, IBM surpasses the state-of-the-art parameter isolation method with a 70\% reduction in the number of parameters within sub-networks and an 80\% decrease in training time.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
CIFAR-100 AlexNet - 300 EpochIBMAccuracy82.69Unverified
CIFAR-100 ResNet-18 - 300 EpochsIBMAccuracy88.15Unverified
MiniImageNet ResNet-18 - 300 EpochsIBMAccuracy53.9Unverified
TinyImageNet ResNet-18 - 300 EpochsIBMAccuracy52.38Unverified

Reproductions