SOTAVerified

Reaching Human-level Performance in Automatic Grammatical Error Correction: An Empirical Study

2018-07-03Code Available0· sign in to hype

Tao Ge, Furu Wei, Ming Zhou

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Neural sequence-to-sequence (seq2seq) approaches have proven to be successful in grammatical error correction (GEC). Based on the seq2seq framework, we propose a novel fluency boost learning and inference mechanism. Fluency boosting learning generates diverse error-corrected sentence pairs during training, enabling the error correction model to learn how to improve a sentence's fluency from more instances, while fluency boosting inference allows the model to correct a sentence incrementally with multiple inference steps. Combining fluency boost learning and inference with convolutional seq2seq models, our approach achieves the state-of-the-art performance: 75.72 (F_0.5) on CoNLL-2014 10 annotation dataset and 62.42 (GLEU) on JFLEG test set respectively, becoming the first GEC system that reaches human-level performance (72.58 for CoNLL and 62.37 for JFLEG) on both of the benchmarks.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
UnrestrictedCNN Seq2Seq + Fluency BoostF0.561.34Unverified
UnrestrictedCNN Seq2Seq + Fluency Boost and inferenceGLEU62.37Unverified

Reproductions