Bidirectional Attention Flow for Machine Comprehension
Minjoon Seo, Aniruddha Kembhavi, Ali Farhadi, Hannaneh Hajishirzi
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/allenai/bi-att-flowOfficialtf★ 0
- github.com/baidu/DuReadertf★ 1,167
- github.com/distractor-generation/dg_surveynone★ 4
- github.com/yugrocks/Question-Answering-with-BIDAFtf★ 0
- github.com/rajatgermany/qa-nlppytorch★ 0
- github.com/GauthierDmn/question_answeringpytorch★ 0
- github.com/WarruzuEndo/BiDAF_mindsporemindspore★ 0
- github.com/xiaobaicxy/Bidaf_SQuAD_MC_Pytorchpytorch★ 0
- github.com/surekhamedapati/NLPA_NEOpytorch★ 0
- github.com/techit-limtiyakul/bidirectional-attention-flowpytorch★ 0
Abstract
Machine comprehension (MC), answering a query about a given context paragraph, requires modeling complex interactions between the context and the query. Recently, attention mechanisms have been successfully extended to MC. Typically these methods use attention to focus on a small portion of the context and summarize it with a fixed-size vector, couple attentions temporally, and/or often form a uni-directional attention. In this paper we introduce the Bi-Directional Attention Flow (BIDAF) network, a multi-stage hierarchical process that represents the context at different levels of granularity and uses bi-directional attention flow mechanism to obtain a query-aware context representation without early summarization. Our experimental evaluations show that our model achieves the state-of-the-art results in Stanford Question Answering Dataset (SQuAD) and CNN/DailyMail cloze test.
Tasks
Benchmark Results
| Dataset | Model | Metric | Claimed | Verified | Status |
|---|---|---|---|---|---|
| QUASAR | BiDAF | EM (Quasar-T) | 25.9 | — | Unverified |