SOTAVerified

Guiding Frame-Level CTC Alignments Using Self-knowledge Distillation

2024-06-12Code Available0· sign in to hype

Eungbeom Kim, Hantae Kim, Kyogu Lee

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Transformer encoder with connectionist temporal classification (CTC) framework is widely used for automatic speech recognition (ASR). However, knowledge distillation (KD) for ASR displays a problem of disagreement between teacher-student models in frame-level alignment which ultimately hinders it from improving the student model's performance. In order to resolve this problem, this paper introduces a self-knowledge distillation (SKD) method that guides the frame-level alignment during the training time. In contrast to the conventional method using separate teacher and student models, this study introduces a simple and effective method sharing encoder layers and applying the sub-model as the student model. Overall, our approach is effective in improving both the resource efficiency as well as performance. We also conducted an experimental analysis of the spike timings to illustrate that the proposed method improves performance by reducing the alignment disagreement.

Tasks

Reproductions