SOTAVerified

Gesture Generation

Generation of gestures, as a sequence of 3d poses

Papers

Showing 150 of 107 papers

TitleStatusHype
DeepGesture: A conversational gesture synthesis system based on emotions and semanticsCode0
Intentional Gesture: Deliver Your Intentions with Gestures for SpeechCode1
M3G: Multi-Granular Gesture Generator for Audio-Driven Full-Body Human Motion Synthesis0
Inter-Diffusion Generation Model of Speakers and Listeners for Effective Communication0
Co^3Gesture: Towards Coherent Concurrent Co-speech 3D Gesture Generation with Interactive Diffusion0
EchoMask: Speech-Queried Attention-based Mask Modeling for Holistic Co-Speech Motion Generation0
EasyGenNet: An Efficient Framework for Audio-Driven Gesture Video Generation Based on Diffusion Model0
Audio-driven Gesture Generation via Deviation Feature in the Latent Space0
SARGes: Semantically Aligned Reliable Gesture Generation via Intent Chain0
DIDiffGes: Decoupled Semi-Implicit Diffusion Models for Real-time Gesture Generation from Speech0
MAG: Multi-Modal Aligned Autoregressive Co-Speech Gesture Generation without Vector Quantization0
Large Language Models for Virtual Human Gesture Selection0
Streaming Generation of Co-Speech Gestures via Accelerated Rolling Diffusion0
HOP: Heterogeneous Topology-based Multimodal Entanglement for Co-Speech Gesture Generation0
VarGes: Improving Variation in Co-Speech 3D Gesture Generation via StyleCLIPSCode0
Contextual Gesture: Co-Speech Gesture Video Generation through Context-aware Gesture Representation0
GestureLSM: Latent Shortcut based Co-Speech Gesture Generation with Spatial-Temporal ModelingCode2
EMO2: End-Effector Guided Audio-Driven Avatar Video Generation0
Co-Speech Gesture Video Generation with Implicit Motion-Audio Entanglement0
SemTalk: Holistic Co-speech Motion Generation with Frame-level Semantic Emphasis0
The Language of Motion: Unifying Verbal and Non-verbal Language of 3D Human Motion0
Retrieving Semantics from the Deep: an RAG Solution for Gesture SynthesisCode2
DiM-Gestor: Co-Speech Gesture Generation with Adaptive Layer Normalization Mamba-20
Conditional GAN for Enhancing Diffusion Models in Efficient and Authentic Global Gesture Generation from Audios0
Large Body Language Models0
Emphasizing Semantic Consistency of Salient Posture for Speech-Driven Gesture Generation0
ExpGest: Expressive Speaker Generation Using Diffusion Model and Hybrid Audio-Text Guidance0
Towards a GENEA Leaderboard -- an Extended, Living Benchmark for Evaluating and Advancing Conversational Motion Synthesis0
LLM Gesticulator: Leveraging Large Language Models for Scalable and Controllable Co-Speech Gesture Synthesis0
Enabling Synergistic Full-Body Control in Prompt-Based Co-Speech Motion GenerationCode0
MM-Conv: A Multi-modal Conversational Dataset for Virtual Humans0
2D or not 2D: How Does the Dimensionality of Gesture Representation Affect 3D Co-Speech Gesture Generation?0
Incorporating Spatial Awareness in Data-Driven Gesture Generation for Virtual Agents0
MDT-A2G: Exploring Masked Diffusion Transformers for Co-Speech Gesture Generation0
DiM-Gesture: Co-Speech Gesture Generation with Adaptive Layer Normalization Mamba-2 framework0
MotionCraft: Crafting Whole-Body Motion with Plug-and-Play Multimodal ControlsCode2
Investigating the impact of 2D gesture representation on co-speech gesture generation0
AMUSE: Emotional Speech-driven 3D Body Animation via Disentangled Latent DiffusionCode2
CoCoGesture: Toward Coherent Co-speech 3D Gesture Generation in the Wild0
LLAniMAtion: LLAMA Driven Gesture Animation0
Bridge to Non-Barrier Communication: Gloss-Prompted Fine-grained Cued Speech Gesture Generation with Diffusion Model0
ConvoFusion: Multi-Modal Conversational Diffusion for Co-Speech Gesture SynthesisCode1
Speech-driven Personalized Gesture Synthetics: Harnessing Automatic Fuzzy Feature Inference0
MambaTalk: Efficient Holistic Gesture Synthesis with Selective State Space ModelsCode2
DiffSHEG: A Diffusion-Based Approach for Real-Time Speech-driven Holistic 3D Expression and Gesture Generation0
Freetalker: Controllable Speech and Text-Driven Gesture Generation Based on Diffusion Models for Enhanced Speaker Naturalness0
EMAGE: Towards Unified Holistic Co-Speech Gesture Generation via Expressive Masked Audio Gesture ModelingCode3
Chain of Generation: Multi-Modal Gesture Synthesis via Cascaded Conditional Control0
Emotional Speech-driven 3D Body Animation via Disentangled Latent DiffusionCode1
Weakly-Supervised Emotion Transition Learning for Diverse 3D Co-speech Gesture Generation0
Show:102550
← PrevPage 1 of 3Next →

No leaderboard results yet.