SOTAVerified

Robust Speech Recognition via Large-Scale Weak Supervision

2022-12-06Preprint 2022Code Available8· sign in to hype

Alec Radford, Jong Wook Kim, Tao Xu, Greg Brockman, Christine McLeavey, Ilya Sutskever

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

We study the capabilities of speech processing systems trained simply to predict large amounts of transcripts of audio on the internet. When scaled to 680,000 hours of multilingual and multitask supervision, the resulting models generalize well to standard benchmarks and are often competitive with prior fully supervised results but in a zero-shot transfer setting without the need for any fine-tuning. When compared to humans, the models approach their accuracy and robustness. We are releasing models and inference code to serve as a foundation for further work on robust speech processing.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
Common Voice EnglishWhisper (Large v2)Word Error Rate (WER)9.4Unverified
Common Voice FrenchWhisper (Large v2)Test WER13.9Unverified
Common Voice GermanWhisper (Large v2)Test WER6.4Unverified
Common Voice ItalianWhisper (Large v2)Test WER7.1Unverified
Common Voice JapaneseWhisper (Large v2)Test WER9.1Unverified
Common Voice RussianWhisper (Large v2)Test WER7.1Unverified
Common Voice SpanishWhisper (Large v2)Test WER5.6Unverified

Reproductions