SOTAVerified

Text-To-Speech Synthesis

Text-To-Speech Synthesis is a machine learning task that involves converting written text into spoken words. The goal is to generate synthetic speech that sounds natural and resembles human speech as closely as possible.

Papers

Showing 201225 of 332 papers

TitleStatusHype
Voice Conversion by Cascading Automatic Speech Recognition and Text-to-Speech Synthesis with Prosody Transfer0
Multi-speaker Multi-style Text-to-speech Synthesis With Single-speaker Single-style Training Data Scenarios0
Multi-speaker Text-to-speech Synthesis Using Deep Gaussian Processes0
Multi-Stage Deep Transfer Learning for EmIoT-enabled Human-Computer Interaction0
Multi-step Natural Language Understanding0
Grad-StyleSpeech: Any-speaker Adaptive Text-to-Speech Synthesis with Diffusion Models0
An Overview of Affective Speech Synthesis and Conversion in the Deep Learning Era0
Neural Harmonic-plus-Noise Waveform Model with Trainable Maximum Voice Frequency for Text-to-Speech Synthesis0
Neural Models of Text Normalization for Speech Applications0
Neural Speech Synthesis in German0
A Novel Data Augmentation Approach for Automatic Speaking Assessment on Opinion Expressions0
Neural Text Normalization with Subword Units0
Neural Text-to-Speech Synthesis for an Under-Resourced Language in a Diglossic Environment: the Case of Gascon Occitan0
Noise-robust zero-shot text-to-speech synthesis conditioned on self-supervised speech-representation model with adapters0
Normalization of Lithuanian Text Using Regular Expressions0
Normalization of Non-Standard Words in Croatian Texts0
Normalizing Text using Language Modelling based on Phonetics and String Similarity0
Open-Source Boundary-Annotated Corpus for Arabic Speech and Language Processing0
VQTTS: High-Fidelity Text-to-Speech Synthesis with Self-Supervised VQ Acoustic Feature0
An objective evaluation of the effects of recording conditions and speaker characteristics in multi-speaker deep neural speech synthesis0
An In-depth Analysis of the Effect of Text Normalization in Social Media0
Parallel WaveNet conditioned on VAE latent vectors0
ParrotTTS: Text-to-Speech synthesis by exploiting self-supervised representations0
Phonetic Enhanced Language Modeling for Text-to-Speech Synthesis0
PnG BERT: Augmented BERT on Phonemes and Graphemes for Neural TTS0
Show:102550
← PrevPage 9 of 14Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1NaturalSpeechAudio Quality MOS4.56Unverified
2VITSAudio Quality MOS4.43Unverified
3Grad-TTS + HiFiGAN (1000 steps)Audio Quality MOS4.37Unverified
4FastSpeech 2 + HiFiGANAudio Quality MOS4.34Unverified
5Glow-TTS + HiFiGANAudio Quality MOS4.34Unverified
6FastSpeech 2 + HiFiGANAudio Quality MOS4.32Unverified
7FastDiff (4 steps)Audio Quality MOS4.28Unverified
8FastDiff-TTSAudio Quality MOS4.03Unverified
9Transformer TTS (Mel + WaveGlow)Audio Quality MOS3.88Unverified
10FastSpeech (Mel + WaveGlow)Audio Quality MOS3.84Unverified
#ModelMetricClaimedVerifiedStatus
1Mia10-keyword Speech Commands dataset16Unverified
#ModelMetricClaimedVerifiedStatus
1Token-Level Ensemble DistillationPhoneme Error Rate4.6Unverified
#ModelMetricClaimedVerifiedStatus
1Tacotron 2Mean Opinion Score3.74Unverified
#ModelMetricClaimedVerifiedStatus
1Tacotron 2Mean Opinion Score3.49Unverified
#ModelMetricClaimedVerifiedStatus
1Match-TTSGMOS3.7Unverified