SOTAVerified

Contrastive Learning-Enhanced Nearest Neighbor Mechanism for Multi-Label Text Classification

2022-05-01ACL 2022Unverified0· sign in to hype

Xi’ao Su, Ran Wang, Xinyu Dai

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

Multi-Label Text Classification (MLTC) is a fundamental and challenging task in natural language processing. Previous studies mainly focus on learning text representation and modeling label correlation but neglect the rich knowledge from the existing similar instances when predicting labels of a specific text. To make up for this oversight, we propose a k nearest neighbor (kNN) mechanism which retrieves several neighbor instances and interpolates the model output with their labels. Moreover, we design a multi-label contrastive learning objective that makes the model aware of the kNN classification process and improves the quality of the retrieved neighbors while inference. Extensive experiments show that our method can bring consistent and significant performance improvement to multiple MLTC models including the state-of-the-art pretrained and non-pretrained ones.

Tasks

Reproductions