SOTAVerified

Relying on Discourse Analysis to Answer Complex Questions by Neural Machine Reading Comprehension

2021-09-01RANLP 2021Unverified0· sign in to hype

Boris Galitsky, Dmitry Ilvovsky, Elizaveta Goncharova

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

Machine reading comprehension (MRC) is one of the most challenging tasks in natural language processing domain. Recent state-of-the-art results for MRC have been achieved with the pre-trained language models, such as BERT and its modifications. Despite the high performance of these models, they still suffer from the inability to retrieve correct answers from the detailed and lengthy passages. In this work, we introduce a novel scheme for incorporating the discourse structure of the text into a self-attention network, and, thus, enrich the embedding obtained from the standard BERT encoder with the additional linguistic knowledge. We also investigate the influence of different types of linguistic information on the model’s ability to answer complex questions that require deep understanding of the whole text. Experiments performed on the SQuAD benchmark and more complex question answering datasets have shown that linguistic enhancing boosts the performance of the standard BERT model significantly.

Tasks

Reproductions