SOTAVerified

UnitedQA: A Hybrid Approach for Open Domain Question Answering

2021-01-01ACL 2021Unverified0· sign in to hype

Hao Cheng, Yelong Shen, Xiaodong Liu, Pengcheng He, Weizhu Chen, Jianfeng Gao

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

To date, most of recent work under the retrieval-reader framework for open-domain QA focuses on either extractive or generative reader exclusively. In this paper, we study a hybrid approach for leveraging the strengths of both models. We apply novel techniques to enhance both extractive and generative readers built upon recent pretrained neural language models, and find that proper training methods can provide large improvement over previous state-of-the-art models. We demonstrate that a simple hybrid approach by combining answers from both readers can efficiently take advantages of extractive and generative answer inference strategies and outperforms single models as well as homogeneous ensembles. Our approach outperforms previous state-of-the-art models by 3.3 and 2.7 points in exact match on NaturalQuestions and TriviaQA respectively.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
Natural QuestionsUnitedQA (Hybrid)Exact Match54.7Unverified
TriviaQAUnitedQA (Hybrid)Exact Match70.5Unverified

Reproductions