SOTAVerified

BERTweet: A pre-trained language model for English Tweets

2020-05-20EMNLP 2020Code Available1· sign in to hype

Dat Quoc Nguyen, Thanh Vu, Anh Tuan Nguyen

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

We present BERTweet, the first public large-scale pre-trained language model for English Tweets. Our BERTweet, having the same architecture as BERT-base (Devlin et al., 2019), is trained using the RoBERTa pre-training procedure (Liu et al., 2019). Experiments show that BERTweet outperforms strong baselines RoBERTa-base and XLM-R-base (Conneau et al., 2020), producing better performance results than the previous state-of-the-art models on three Tweet NLP tasks: Part-of-speech tagging, Named-entity recognition and text classification. We release BERTweet under the MIT License to facilitate future research and applications on Tweet data. Our BERTweet is available at https://github.com/VinAIResearch/BERTweet

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
WNUT 2016BERTweetF152.1Unverified
WNUT 2017BERTweetF156.5Unverified

Reproductions