SyncVSR: Data-Efficient Visual Speech Recognition with End-to-End Crossmodal Audio Token Synchronization
Young Jin Ahn, Jungwoo Park, Sangha Park, Jonghyun Choi, Kee-Eung Kim
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/KAIST-AILab/SyncVSROfficialjax★ 60
Abstract
Visual Speech Recognition (VSR) stands at the intersection of computer vision and speech recognition, aiming to interpret spoken content from visual cues. A prominent challenge in VSR is the presence of homophenes-visually similar lip gestures that represent different phonemes. Prior approaches have sought to distinguish fine-grained visemes by aligning visual and auditory semantics, but often fell short of full synchronization. To address this, we present SyncVSR, an end-to-end learning framework that leverages quantized audio for frame-level crossmodal supervision. By integrating a projection layer that synchronizes visual representation with acoustic data, our encoder learns to generate discrete audio tokens from a video sequence in a non-autoregressive manner. SyncVSR shows versatility across tasks, languages, and modalities at the cost of a forward pass. Our empirical evaluations show that it not only achieves state-of-the-art results but also reduces data usage by up to ninefold.
Tasks
Benchmark Results
| Dataset | Model | Metric | Claimed | Verified | Status |
|---|---|---|---|---|---|
| CAS-VSR-W1k (LRW-1000) | SyncVSR (Word Boundary) | Top-1 Accuracy | 58.2 | — | Unverified |
| Lip Reading in the Wild | SyncVSR | Top-1 Accuracy | 93.2 | — | Unverified |
| Lip Reading in the Wild | SyncVSR (Word Boundary) | Top-1 Accuracy | 95 | — | Unverified |
| LRS2 | SyncVSR | Word Error Rate (WER) | 28.9 | — | Unverified |
| LRS2 | SyncVSR | Word Error Rate (WER) | 16.5 | — | Unverified |
| LRS3-TED | SyncVSR | Word Error Rate (WER) | 21.5 | — | Unverified |
| LRS3-TED | SyncVSR | Word Error Rate (WER) | 31.2 | — | Unverified |