SOTAVerified

A Study of Residual Adapters for Multi-Domain Neural Machine Translation

2020-11-01WMT (EMNLP) 2020Code Available0· sign in to hype

Minh Quang Pham, Josep Maria Crego, François Yvon, Jean Senellart

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Domain adaptation is an old and vexing problem for machine translation systems. The most common approach and successful to supervised adaptation is to fine-tune a baseline system with in-domain parallel data. Standard fine-tuning however modifies all the network parameters, which makes this approach computationally costly and prone to overfitting. A recent, lightweight approach, instead augments a baseline model with supplementary (small) adapter layers, keeping the rest of the mode unchanged. This has the additional merit to leave the baseline model intact, and adaptable to multiple domains. In this paper, we conduct a thorough analysis of the adapter model in the context of a multidomain machine translation task. We contrast multiple implementations of this idea on two language pairs. Our main conclusions are that residual adapters provide a fast and cheap method for supervised multi-domain adaptation; our two variants prove as effective as the original adapter model, and open perspective to also make adapted models more robust to label domain errors.

Tasks

Reproductions