Wav2Letter: an End-to-End ConvNet-based Speech Recognition System
2016-09-11arXiv 2016Code Available1· sign in to hype
Ronan Collobert, Christian Puhrsch, Gabriel Synnaeve
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/mailong25/vietnamese-speech-recognitionpytorch★ 379
- github.com/ashwin9999/Capstone-Speech-to-SQLtf★ 0
- github.com/silversparro/wav2letter.pytorchpytorch★ 0
- github.com/MrMao/wav2lettertorch★ 0
- github.com/ashwin9999/speech-recognition-CNNtf★ 0
- github.com/CorrelAid/codingchallenge1020_team1tf★ 0
- github.com/JuliusKunze/speechlesstf★ 0
- github.com/eric-erki/wav2lettertorch★ 0
Abstract
This paper presents a simple end-to-end model for speech recognition, combining a convolutional network based acoustic model and a graph decoding. It is trained to output letters, with transcribed speech, without the need for force alignment of phonemes. We introduce an automatic segmentation criterion for training from sequence annotation without alignment that is on par with CTC while being simpler. We show competitive results in word error rate on the Librispeech corpus with MFCC features, and promising results from raw waveform.