SOTAVerified

Faster, Simpler and More Accurate Hybrid ASR Systems Using Wordpieces

2020-05-19Unverified0· sign in to hype

Frank Zhang, Yongqiang Wang, Xiaohui Zhang, Chunxi Liu, Yatharth Saraf, Geoffrey Zweig

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

In this work, we first show that on the widely used LibriSpeech benchmark, our transformer-based context-dependent connectionist temporal classification (CTC) system produces state-of-the-art results. We then show that using wordpieces as modeling units combined with CTC training, we can greatly simplify the engineering pipeline compared to conventional frame-based cross-entropy training by excluding all the GMM bootstrapping, decision tree building and force alignment steps, while still achieving very competitive word-error-rate. Additionally, using wordpieces as modeling units can significantly improve runtime efficiency since we can use larger stride without losing accuracy. We further confirm these findings on two internal VideoASR datasets: German, which is similar to English as a fusional language, and Turkish, which is an agglutinative language.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
LibriSpeech test-cleanCTC + Transformer LM rescoringWord Error Rate (WER)2.1Unverified
LibriSpeech test-otherCTC + Transformer LM rescoringWord Error Rate (WER)4.2Unverified

Reproductions