SOTAVerified

Talking Face Generation

Talking face generation aims to synthesize a sequence of face images that correspond to given speech semantics

( Image credit: Talking Face Generation by Adversarially Disentangled Audio-Visual Representation )

Papers

Showing 3140 of 110 papers

TitleStatusHype
Flow-Guided One-Shot Talking Face Generation With a High-Resolution Audio-Visual DatasetCode1
Text2Video: Text-driven Talking-head Video Synthesis with Personalized Phoneme-Pose DictionaryCode1
Pose-Controllable Talking Face Generation by Implicitly Modularized Audio-Visual RepresentationCode1
Speech Driven Talking Face Generation from a Single Image and an Emotion ConditionCode1
MEAD: A Large-scale Audio-visual Dataset for Emotional Talking-face GenerationCode1
DisentTalk: Cross-lingual Talking Face Generation via Semantic Disentangled Diffusion Model0
UniSync: A Unified Framework for Audio-Visual Synchronization0
PC-Talk: Precise Facial Animation Control for Audio-Driven Talking Face Generation0
Playmate: Flexible Control of Portrait Animation via 3D-Implicit Space Guided Diffusion0
GLCF: A Global-Local Multimodal Coherence Analysis Framework for Talking Face Generation Detection0
Show:102550
← PrevPage 4 of 11Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1EmoGenEmoAcc83.2Unverified
#ModelMetricClaimedVerifiedStatus
1LipGANLMD0.6Unverified