SOTAVerified

End-to-End Speech Translation with Pre-trained Models and Adapters: UPC at IWSLT 2021

2021-05-10ACL (IWSLT) 2021Code Available1· sign in to hype

Gerard I. Gállego, Ioannis Tsiamas, Carlos Escolano, José A. R. Fonollosa, Marta R. Costa-jussà

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

This paper describes the submission to the IWSLT 2021 offline speech translation task by the UPC Machine Translation group. The task consists of building a system capable of translating English audio recordings extracted from TED talks into German text. Submitted systems can be either cascade or end-to-end and use a custom or given segmentation. Our submission is an end-to-end speech translation system, which combines pre-trained models (Wav2Vec 2.0 and mBART) with coupling modules between the encoder and decoder, and uses an efficient fine-tuning technique, which trains only 20% of its total parameters. We show that adding an Adapter to the system and pre-training it, can increase the convergence speed and the final result, with which we achieve a BLEU score of 27.3 on the MuST-C test set. Our final model is an ensemble that obtains 28.22 BLEU score on the same set. Our submission also uses a custom segmentation algorithm that employs pre-trained Wav2Vec 2.0 for identifying periods of untranscribable text and can bring improvements of 2.5 to 3 BLEU score on the IWSLT 2019 test set, as compared to the result with the given segmentation.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
MuST-C EN->DEWav2Vec2.0+mBART+AdaptorsCase-sensitive sacreBLEU28.22Unverified

Reproductions