SOTAVerified

TR at SemEval-2020 Task 4: Exploring the Limits of Language-model-based Common Sense Validation

2020-12-01SEMEVALUnverified0· sign in to hype

Don Teo

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

In this paper, we present our submission for subtask A of the Common Sense Validation and Explanation (ComVE) shared task. We examine the ability of large-scale pre-trained language models to distinguish commonsense from non-commonsense statements. We also explore the utility of external resources that aim to supplement the world knowledge inherent in such language models, including commonsense knowledge graph embedding models, word concreteness ratings, and text-to-image generation models. We find that such resources provide insignificant gains to the performance of fine-tuned language models. We also provide a qualitative analysis of the limitations of the language model fine-tuned to this task.

Tasks

Reproductions