SOTAVerified

Adapting BERT for Continual Learning of a Sequence of Aspect Sentiment Classification Tasks

2021-12-06NAACL 2021Code Available0· sign in to hype

Zixuan Ke, Hu Xu, Bing Liu

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

This paper studies continual learning (CL) of a sequence of aspect sentiment classification (ASC) tasks. Although some CL techniques have been proposed for document sentiment classification, we are not aware of any CL work on ASC. A CL system that incrementally learns a sequence of ASC tasks should address the following two issues: (1) transfer knowledge learned from previous tasks to the new task to help it learn a better model, and (2) maintain the performance of the models for previous tasks so that they are not forgotten. This paper proposes a novel capsule network based model called B-CL to address these issues. B-CL markedly improves the ASC performance on both the new task and the old tasks via forward and backward knowledge transfer. The effectiveness of B-CL is demonstrated through extensive experiments.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
20Newsgroup (10 tasks)B-CLF1 - macro0.95Unverified
ASC (19 tasks)B-CLF1 - macro0.81Unverified
DSC (10 tasks)B-CLF1 - macro0.77Unverified

Reproductions