SOTAVerified

Multi-split Reversible Transformers Can Enhance Neural Machine Translation

2021-04-01EACL 2021Unverified0· sign in to hype

Yuekai Zhao, Shuchang Zhou, Zhihua Zhang

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

Large-scale transformers have been shown the state-of-the-art on neural machine translation. However, training these increasingly wider and deeper models could be tremendously memory intensive. We reduce the memory burden by employing the idea of reversible networks that a layer's input can be reconstructed from its output. We design three types of multi-split based reversible transformers. We also devise a corresponding backpropagation algorithm, which does not need to store activations for most layers. Furthermore, we present two fine-tuning techniques: splits shuffle and self ensemble, to boost translation accuracy. Specifically, our best models surpass the vanilla transformer by at least 1.4 BLEU points in three datasets. Our large-scale reversible models achieve 30.0 BLEU in WMT'14 En-De and 43.5 BLEU in WMT'14 En-Fr, beating several very strong baselines with less than half of the training memory.

Tasks

Reproductions