SOTAVerified

Enolp musk@SMM4H’22 : Leveraging Pre-trained Language Models for Stance And Premise Classification

2022-10-01SMM4H (COLING) 2022Code Available0· sign in to hype

Millon Das, Archit Mangrulkar, Ishan Manchanda, Manav Kapadnis, Sohan Patnaik

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

This paper covers our approaches for the Social Media Mining for Health (SMM4H) Shared Tasks 2a and 2b. Apart from the baseline architectures, we experiment with Parts of Speech (PoS), dependency parsing, and Tf-Idf features. Additionally, we perform contrastive pretraining on our best models using a supervised contrastive loss function. In both the tasks, we outperformed the mean and median scores and ranked first on the validation set. For stance classification, we achieved an F1-score of 0.636 using the CovidTwitterBERT model, while for premise classification, we achieved an F1-score of 0.664 using BART-base model on test dataset.

Tasks

Reproductions