SOTAVerified

Incremental Domain Adaptation for Neural Machine Translation in Low-Resource Settings

2019-08-01WS 2019Code Available0· sign in to hype

Marimuthu Kalimuthu, Michael Barz, Daniel Sonntag

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

We study the problem of incremental domain adaptation of a generic neural machine translation model with limited resources (e.g., budget and time) for human translations or model training. In this paper, we propose a novel query strategy for selecting ``unlabeled'' samples from a new domain based on sentence embeddings for Arabic. We accelerate the fine-tuning process of the generic model to the target domain. Specifically, our approach estimates the informativeness of instances from the target domain by comparing the distance of their sentence embeddings to embeddings from the generic domain. We perform machine translation experiments (Ar-to-En direction) for comparing a random sampling baseline with our new approach, similar to active learning, using two small update sets for simulating the work of human translators. For the prescribed setting we can save more than 50\% of the annotation costs without loss in quality, demonstrating the effectiveness of our approach.

Tasks

Reproductions