SOTAVerified

A Recurrent BERT-based Model for Question Generation

2019-11-01WS 2019Code Available0· sign in to hype

Ying-Hong Chan, Yao-Chung Fan

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

In this study, we investigate the employment of the pre-trained BERT language model to tackle question generation tasks. We introduce three neural architectures built on top of BERT for question generation tasks. The first one is a straightforward BERT employment, which reveals the defects of directly using BERT for text generation. Accordingly, we propose another two models by restructuring our BERT employment into a sequential manner for taking information from previous decoded results. Our models are trained and evaluated on the recent question-answering dataset SQuAD. Experiment results show that our best model yields state-of-the-art performance which advances the BLEU 4 score of the existing best models from 16.85 to 22.17.

Tasks

Reproductions