Towards Neural Phrase-based Machine Translation
Po-Sen Huang, Chong Wang, Sitao Huang, Dengyong Zhou, Li Deng
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/posenhuang/NPMTOfficialIn papertorch★ 0
- github.com/Microsoft/NPMTtorch★ 0
- github.com/ykrmm/ICLR_2020pytorch★ 0
- github.com/ykrmm/TREMBApytorch★ 0
Abstract
In this paper, we present Neural Phrase-based Machine Translation (NPMT). Our method explicitly models the phrase structures in output sequences using Sleep-WAke Networks (SWAN), a recently proposed segmentation-based sequence modeling method. To mitigate the monotonic alignment requirement of SWAN, we introduce a new layer to perform (soft) local reordering of input sequences. Different from existing neural machine translation (NMT) approaches, NPMT does not use attention-based decoding mechanisms. Instead, it directly outputs phrases in a sequential order and can decode in linear time. Our experiments show that NPMT achieves superior performances on IWSLT 2014 German-English/English-German and IWSLT 2015 English-Vietnamese machine translation tasks compared with strong NMT baselines. We also observe that our method produces meaningful phrases in output languages.
Tasks
Benchmark Results
| Dataset | Model | Metric | Claimed | Verified | Status |
|---|---|---|---|---|---|
| IWSLT2014 German-English | Neural PBMT + LM [Huang2018] | BLEU score | 30.08 | — | Unverified |
| IWSLT2015 English-German | NPMT + language model | BLEU score | 25.36 | — | Unverified |
| IWSLT2015 German-English | NPMT + language model | BLEU score | 30.08 | — | Unverified |