SOTAVerified

Zipformer: A faster and better encoder for automatic speech recognition

2023-10-17Code Available0· sign in to hype

Zengwei Yao, Liyong Guo, Xiaoyu Yang, Wei Kang, Fangjun Kuang, Yifan Yang, Zengrui Jin, Long Lin, Daniel Povey

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

The Conformer has become the most popular encoder model for automatic speech recognition (ASR). It adds convolution modules to a transformer to learn both local and global dependencies. In this work we describe a faster, more memory-efficient, and better-performing transformer, called Zipformer. Modeling changes include: 1) a U-Net-like encoder structure where middle stacks operate at lower frame rates; 2) reorganized block structure with more modules, within which we re-use attention weights for efficiency; 3) a modified form of LayerNorm called BiasNorm allows us to retain some length information; 4) new activation functions SwooshR and SwooshL work better than Swish. We also propose a new optimizer, called ScaledAdam, which scales the update by each tensor's current scale to keep the relative change about the same, and also explictly learns the parameter scale. It achieves faster convergence and better performance than Adam. Extensive experiments on LibriSpeech, Aishell-1, and WenetSpeech datasets demonstrate the effectiveness of our proposed Zipformer over other state-of-the-art ASR models. Our code is publicly available at https://github.com/k2-fsa/icefall.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
LibriSpeech test-cleanZipformer+pruned transducer (no external language model)Word Error Rate (WER)2Unverified
LibriSpeech test-otherZipformer+pruned transducer (no external language model)Word Error Rate (WER)4.38Unverified
WenetSpeechZipformer+pruned transducer (no external language model)Character Error Rate (CER)7.29Unverified

Reproductions