SOTAVerified

Clustering Examples in Multi-Dataset Benchmarks with Item Response Theory

2022-05-01insights (ACL) 2022Unverified0· sign in to hype

Pedro Rodriguez, Phu Mon Htut, John Lalor, João Sedoc

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

In natural language processing, multi-dataset benchmarks for common tasks (e.g., SuperGLUE for natural language inference and MRQA for question answering) have risen in importance. Invariably, tasks and individual examples vary in difficulty. Recent analysis methods infer properties of examples such as difficulty. In particular, Item Response Theory (IRT) jointly infers example and model properties from the output of benchmark tasks (i.e., scores for each model-example pair). Therefore, it seems sensible that methods like IRT should be able to detect differences between datasets in a task. This work shows that current IRT models are not as good at identifying differences as we would expect, explain why this is difficult, and outline future directions that incorporate more (textual) signal from examples.

Tasks

Reproductions