SOTAVerified

Towards Modeling the Style of Translators in Neural Machine Translation

2021-06-01NAACL 2021Unverified0· sign in to hype

Yue Wang, Cuong Hoang, Marcello Federico

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

One key ingredient of neural machine translation is the use of large datasets from different domains and resources (e.g. Europarl, TED talks). These datasets contain documents translated by professional translators using different but consistent translation styles. Despite that, the model is usually trained in a way that neither explicitly captures the variety of translation styles present in the data nor translates new data in different and controllable styles. In this work, we investigate methods to augment the state of the art Transformer model with translator information that is available in part of the training data. We show that our style-augmented translation models are able to capture the style variations of translators and to generate translations with different styles on new data. Indeed, the generated variations differ significantly, up to +4.5 BLEU score difference. Despite that, human evaluation confirms that the translations are of the same quality.

Tasks

Reproductions