SOTAVerified

UnifiedQA: Crossing Format Boundaries With a Single QA System

2020-05-02Findings of the Association for Computational LinguisticsCode Available1· sign in to hype

Daniel Khashabi, Sewon Min, Tushar Khot, Ashish Sabharwal, Oyvind Tafjord, Peter Clark, Hannaneh Hajishirzi

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Question answering (QA) tasks have been posed using a variety of formats, such as extractive span selection, multiple choice, etc. This has led to format-specialized models, and even to an implicit division in the QA community. We argue that such boundaries are artificial and perhaps unnecessary, given the reasoning abilities we seek to teach are not governed by the format. As evidence, we use the latest advances in language modeling to build a single pre-trained QA model, UnifiedQA, that performs surprisingly well across 17 QA datasets spanning 4 diverse formats. UnifiedQA performs on par with 9 different models that were trained on individual datasets themselves. Even when faced with 12 unseen datasets of observed formats, UnifiedQA performs surprisingly well, showing strong generalization from its out-of-format training data. Finally, simply fine-tuning this pre-trained QA model into specialized models results in a new state of the art on 6 datasets, establishing UnifiedQA as a strong starting point for building QA systems.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
CommonsenseQABART-large 440M (fine-tuned)Accuracy62.5Unverified
CommonsenseQAUnifiedQA 11B (fine-tuned)Accuracy79.1Unverified
CommonsenseQAT5-XXL 11B (fine-tuned)Accuracy78.1Unverified
CommonsenseQAUnifiedQA 11B (zero-shot)Accuracy76.2Unverified
CommonsenseQAUnifiedQA 440M (fine-tuned)Accuracy64Unverified
WinoGrandeUnifiedQA 11B (fine-tuned)Accuracy89.4Unverified
WinoGrandeUnified QA 406M (fine-tuned)Accuracy73.3Unverified

Reproductions