SOTAVerified

Cluster & Tune: Enhance BERT Performance in Low Resource Text Classification

2021-01-01Unverified0· sign in to hype

Eyal Shnarch, Ariel Gera, Alon Halfon, Lena Dankin, Leshem Choshen, Ranit Aharonov, Noam Slonim

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

In data-constrained cases, the common practice of fine-tuning BERT for a target text classification task is prone to producing poor performance. In such low resources scenarios, we suggest performing an unsupervised classification task prior to fine-tuning on the target task. Specifically, as such an intermediate task, we perform unsupervised clustering, training BERT on predicting the cluster labels. We test this hypothesis on various data sets, and show that this additional classification step can reduce the demand for labeled examples. We further discuss under which conditions this task is helpful and why.

Tasks

Reproductions