SOTAVerified

Noobs at Semeval-2021 Task 4: Masked Language Modeling for abstract answer prediction

2021-08-01SEMEVALUnverified0· sign in to hype

Shikhar Shukla, Sarthak Sarthak, Karm Veer Arya

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

This paper presents the system developed by our team for Semeval 2021 Task 4: Reading Comprehension of Abstract Meaning. The aim of the task was to benchmark the NLP techniques in understanding the abstract concepts present in a passage, and then predict the missing word in a human written summary of the passage. We trained a Roberta-Large model trained with a masked language modeling objective. In cases where this model failed to predict one of the available options, another Roberta-Large model trained as a binary classifier was used to predict correct and incorrect options. We used passage summary generated by Pegasus model and question as inputs. Our best solution was an ensemble of these 2 systems. We achieved an accuracy of 86.22\% on subtask 1 and 87.10\% on subtask 2.

Tasks

Reproductions