UCAS-IIE-NLP at SemEval-2023 Task 12: Enhancing Generalization of Multilingual BERT for Low-resource Sentiment Analysis
Dou Hu, Lingwei Wei, Yaxin Liu, Wei Zhou, Songlin Hu
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/zerohd4869/saclpytorch★ 36
Abstract
This paper describes our system designed for SemEval-2023 Task 12: Sentiment analysis for African languages. The challenge faced by this task is the scarcity of labeled data and linguistic resources in low-resource settings. To alleviate these, we propose a generalized multilingual system SACL-XLMR for sentiment analysis on low-resource languages. Specifically, we design a lexicon-based multilingual BERT to facilitate language adaptation and sentiment-aware representation learning. Besides, we apply a supervised adversarial contrastive learning technique to learn sentiment-spread structured representations and enhance model generalization. Our system achieved competitive results, largely outperforming baselines on both multilingual and zero-shot sentiment classification subtasks. Notably, the system obtained the 1st rank on the zero-shot classification subtask in the official ranking. Extensive experiments demonstrate the effectiveness of our system.
Tasks
Benchmark Results
| Dataset | Model | Metric | Claimed | Verified | Status |
|---|---|---|---|---|---|
| AfriSenti | SACL-XLMR | weighted-F1 score | 0.59 | — | Unverified |
| AfriSenti | AfroXLMR | weighted-F1 score | 0.56 | — | Unverified |
| AfriSenti | AfriBERTa | weighted-F1 score | 0.44 | — | Unverified |
| AfriSenti | XLM-R | weighted-F1 score | 0.4 | — | Unverified |
| AfriSenti | Random | weighted-F1 score | 0.34 | — | Unverified |