SOTAVerified

Multimodal Emotion Recognition

This is a leaderboard for multimodal emotion recognition on the IEMOCAP dataset. The modality abbreviations are A: Acoustic T: Text V: Visual

Please include the modality in the bracket after the model name.

All models must use standard five emotion categories and are evaluated in standard leave-one-session-out (LOSO). See the papers for references.

Papers

Showing 76100 of 180 papers

TitleStatusHype
A Multimodal Emotion Recognition System: Integrating Facial Expressions, Body Movement, Speech, and Spoken Language0
Dynamic Graph Neural ODE Network for Multi-modal Emotion Recognition in Conversation0
CMATH: Cross-Modality Augmented Transformer with Hierarchical Variational Distillation for Multimodal Emotion Recognition in Conversation0
Smile upon the Face but Sadness in the Eyes: Emotion Recognition based on Facial Expressions and Eye Behaviors0
Early Joint Learning of Emotion Information Makes MultiModal Model Understand You Better0
Multimodal Emotion Recognition with Vision-language Prompting and Modality Dropout0
Leveraging Contrastive Learning and Self-Training for Multimodal Emotion Recognition with Limited Labeled SamplesCode0
Video Emotion Open-vocabulary Recognition Based on Multimodal Large Language Model0
Multi Teacher Privileged Knowledge Distillation for Multimodal Expression RecognitionCode0
MicroEmo: Time-Sensitive Multimodal Emotion Recognition with Micro-Expression Dynamics in Video Dialogues0
Masked Graph Learning with Recurrent Alignment for Multimodal Emotion Recognition in Conversation0
Textualized and Feature-based Models for Compound Multimodal Emotion Recognition in the WildCode0
Enhancing Emotion Recognition in Incomplete Data: A Novel Cross-Modal Alignment, Reconstruction, and Refinement Framework0
Multimodal Emotion Recognition based on Facial Expressions, Speech, and EEG0
Inconsistency-Aware Cross-Attention for Audio-Visual Fusion in Dimensional Emotion Recognition0
Empathy Through Multimodality in Conversational Interfaces0
Revisiting Multimodal Emotion Recognition in Conversation from the Perspective of Graph Spectrum0
Dynamic Modality and View Selection for Multimodal Emotion Recognition with Missing Modalities0
Deep CNN with late fusion for realtime multimodal emotion recognition0
Multimodal Emotion Recognition by Fusing Video Semantic in MOOC Learning Scenarios0
UniMEEC: Towards Unified Multimodal Emotion Recognition and Emotion Cause0
Multi-Modal Emotion Recognition by Text, Speech and Video Using Pretrained Transformers0
A Two-Stage Multimodal Emotion Recognition Model Based on Graph Contrastive Learning0
MART: Masked Affective RepresenTation Learning via Masked Temporal Distribution Distillation0
Adversarial Representation with Intra-Modal and Inter-Modal Graph Contrastive Learning for Multimodal Emotion Recognition0
Show:102550
← PrevPage 4 of 8Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1GraphSmileWeighted F186.52Unverified
2JoyfulWeighted F185.7Unverified
3COGMENWeighted F184.5Unverified
4DANNAccuracy82.7Unverified
5MMERAccuracy81.7Unverified
6PATHOSnet v2Accuracy80.4Unverified
7Self-attention weight correction (A+T)Accuracy76.8Unverified
8CHFusionAccuracy76.5Unverified
9bc-LSTMWeighted F174.1Unverified
10Audio + Text (Stage III)F170.5Unverified
#ModelMetricClaimedVerifiedStatus
1GraphSmileWeighted F166.71Unverified
2Audio + Text (Stage III)Weighted F165.8Unverified
3JoyfulWeighted F161.77Unverified
#ModelMetricClaimedVerifiedStatus
1GraphSmileWeighted F172.81Unverified
2JoyfulWeighted F170.5Unverified
#ModelMetricClaimedVerifiedStatus
1GraphSmileWeighted F144.93Unverified
#ModelMetricClaimedVerifiedStatus
1GraphSmileWeighted F166.73Unverified
#ModelMetricClaimedVerifiedStatus
1SMPLify-Xv2v error52.9Unverified
#ModelMetricClaimedVerifiedStatus
1GraphSmileWeighted F174.31Unverified