SOTAVerified

The AFRL IWSLT 2020 Systems: Work-From-Home Edition

2020-07-01WS 2020Unverified0· sign in to hype

Brian Ore, Eric Hansen, Tim Anderson, Jeremy Gwinnup

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

This report summarizes the Air Force Research Laboratory (AFRL) submission to the offline spoken language translation (SLT) task as part of the IWSLT 2020 evaluation campaign. As in previous years, we chose to adopt the cascade approach of using separate systems to perform speech activity detection, automatic speech recognition, sentence segmentation, and machine translation. All systems were neural based, including a fully-connected neural network for speech activity detection, a Kaldi factorized time delay neural network with recurrent neural network (RNN) language model rescoring for speech recognition, a bidirectional RNN with attention mechanism for sentence segmentation, and transformer networks trained with OpenNMT and Marian for machine translation. Our primary submission yielded BLEU scores of 21.28 on tst2019 and 23.33 on tst2020.

Tasks

Reproductions