SOTAVerified

Text-To-Speech Synthesis

Text-To-Speech Synthesis is a machine learning task that involves converting written text into spoken words. The goal is to generate synthetic speech that sounds natural and resembles human speech as closely as possible.

Papers

Showing 101125 of 332 papers

TitleStatusHype
Mlphon: A Multifunctional Grapheme-Phoneme Conversion Tool Using Finite State TransducersCode0
Independent and automatic evaluation of acoustic-to-articulatory inversion modelsCode0
Investigation of enhanced Tacotron text-to-speech synthesis systems with self-attention for pitch accent languageCode0
Speech Synthesis from Text and Ultrasound Tongue Image-based Articulatory InputCode0
Creating New Language and Voice Components for the Updated MaryTTS Text-to-Speech Synthesis Platform0
Controllable Prosody Generation With Partial Inputs0
A unified sequence-to-sequence front-end model for Mandarin text-to-speech synthesis0
Controllable neural text-to-speech synthesis using intuitive prosodic features0
Controllable Accented Text-to-Speech Synthesis0
A unified front-end framework for English text-to-speech synthesis0
An Overview of Affective Speech Synthesis and Conversion in the Deep Learning Era0
Continual Speaker Adaptation for Text-to-Speech Synthesis0
Full-text Error Correction for Chinese Speech Recognition with Large Language Model0
FMSD-TTS: Few-shot Multi-Speaker Multi-Dialect Text-to-Speech Synthesis for Ü-Tsang, Amdo and Kham Speech Dataset Generation0
Conditioning Sequence-to-sequence Networks with Learned Activations0
A Unified Framework for Collecting Text-to-Speech Synthesis Datasets for 22 Indian Languages0
FLY-TTS: Fast, Lightweight and High-Quality End-to-End Text-to-Speech Synthesis0
Generative adversarial network-based glottal waveform model for statistical parametric speech synthesis0
Flavored Tacotron: Conditional Learning for Prosodic-linguistic Features0
Fine-grained Style Modeling, Transfer and Prediction in Text-to-Speech Synthesis via Phone-Level Content-Style Disentanglement0
Generative Pre-training for Speech with Flow Matching0
Generative Semantic Communication for Text-to-Speech Synthesis0
Comparing normalizing flows and diffusion models for prosody and acoustic modelling in text-to-speech0
Augmenting Images for ASR and TTS through Single-loop and Dual-loop Multimodal Chain Framework0
A Novel Data Augmentation Approach for Automatic Speaking Assessment on Opinion Expressions0
Show:102550
← PrevPage 5 of 14Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1NaturalSpeechAudio Quality MOS4.56Unverified
2VITSAudio Quality MOS4.43Unverified
3Grad-TTS + HiFiGAN (1000 steps)Audio Quality MOS4.37Unverified
4FastSpeech 2 + HiFiGANAudio Quality MOS4.34Unverified
5Glow-TTS + HiFiGANAudio Quality MOS4.34Unverified
6FastSpeech 2 + HiFiGANAudio Quality MOS4.32Unverified
7FastDiff (4 steps)Audio Quality MOS4.28Unverified
8FastDiff-TTSAudio Quality MOS4.03Unverified
9Transformer TTS (Mel + WaveGlow)Audio Quality MOS3.88Unverified
10FastSpeech (Mel + WaveGlow)Audio Quality MOS3.84Unverified
#ModelMetricClaimedVerifiedStatus
1Mia10-keyword Speech Commands dataset16Unverified
#ModelMetricClaimedVerifiedStatus
1Token-Level Ensemble DistillationPhoneme Error Rate4.6Unverified
#ModelMetricClaimedVerifiedStatus
1Tacotron 2Mean Opinion Score3.74Unverified
#ModelMetricClaimedVerifiedStatus
1Tacotron 2Mean Opinion Score3.49Unverified
#ModelMetricClaimedVerifiedStatus
1Match-TTSGMOS3.7Unverified