SOTAVerified

Non-Autoregressive Neural Machine Translation

2017-11-07ICLR 2018Code Available0· sign in to hype

Jiatao Gu, James Bradbury, Caiming Xiong, Victor O. K. Li, Richard Socher

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Existing approaches to neural machine translation condition each output word on previously generated outputs. We introduce a model that avoids this autoregressive property and produces its outputs in parallel, allowing an order of magnitude lower latency during inference. Through knowledge distillation, the use of input token fertilities as a latent variable, and policy gradient fine-tuning, we achieve this at a cost of as little as 2.0 BLEU points relative to the autoregressive Transformer network used as a teacher. We demonstrate substantial cumulative improvements associated with each of the three aspects of our training strategy, and validate our approach on IWSLT 2016 English-German and two WMT language pairs. By sampling fertilities in parallel at inference time, our non-autoregressive model achieves near-state-of-the-art performance of 29.8 BLEU on WMT 2016 English-Romanian.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
IWSLT2015 English-GermanNAT +FT + NPDBLEU score28.16Unverified
WMT2014 English-GermanNAT +FT + NPDBLEU score19.17Unverified
WMT2014 German-EnglishNAT +FT + NPDBLEU score23.2Unverified
WMT2016 English-RomanianNAT +FT + NPDBLEU score29.79Unverified
WMT2016 Romanian-EnglishNAT +FT + NPDBLEU score31.44Unverified

Reproductions