Lightweight Adapter Tuning for Multilingual Speech Translation
Hang Le, Juan Pino, Changhan Wang, Jiatao Gu, Didier Schwab, Laurent Besacier
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/formiel/fairseqOfficialIn paperpytorch★ 18
- github.com/formiel/fairseq/blob/master/examples/speech_to_text/docs/adapters.mdOfficialpytorch★ 0
Abstract
Adapter modules were recently introduced as an efficient alternative to fine-tuning in NLP. Adapter tuning consists in freezing pretrained parameters of a model and injecting lightweight modules between layers, resulting in the addition of only a small number of task-specific trainable parameters. While adapter tuning was investigated for multilingual neural machine translation, this paper proposes a comprehensive analysis of adapters for multilingual speech translation (ST). Starting from different pre-trained models (a multilingual ST trained on parallel data or a multilingual BART (mBART) trained on non-parallel multilingual data), we show that adapters can be used to: (a) efficiently specialize ST to specific language pairs with a low extra cost in terms of parameters, and (b) transfer from an automatic speech recognition (ASR) task and an mBART pre-trained model to a multilingual ST task. Experiments show that adapter tuning offer competitive results to full fine-tuning, while being much more parameter-efficient.
Tasks
Benchmark Results
| Dataset | Model | Metric | Claimed | Verified | Status |
|---|---|---|---|---|---|
| MuST-C | Transformer with Adapters | SacreBLEU | 26.61 | — | Unverified |
| MuST-C EN->DE | Transformer with Adapters | Case-sensitive sacreBLEU | 24.63 | — | Unverified |
| MuST-C EN->ES | Transformer with Adapters | Case-sensitive sacreBLEU | 28.73 | — | Unverified |