SOTAVerified

Identifying inherent disagreement in natural language inference

2021-06-01NAACL 2021Code Available0· sign in to hype

Xinliang Frederick Zhang, Marie-Catherine de Marneffe

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Natural language inference (NLI) is the task of determining whether a piece of text is entailed, contradicted by or unrelated to another piece of text. In this paper, we investigate how to tease systematic inferences (i.e., items for which people agree on the NLI label) apart from disagreement items (i.e., items which lead to different annotations), which most prior work has overlooked. To distinguish systematic inferences from disagreement items, we propose Artificial Annotators (AAs) to simulate the uncertainty in the annotation process by capturing the modes in annotations. Results on the CommitmentBank, a corpus of naturally occurring discourses in English, confirm that our approach performs statistically significantly better than all baselines. We further show that AAs learn linguistic patterns and context-dependent reasoning.

Tasks

Reproductions