SOTAVerified

Multimodal Emotion Recognition

This is a leaderboard for multimodal emotion recognition on the IEMOCAP dataset. The modality abbreviations are A: Acoustic T: Text V: Visual

Please include the modality in the bracket after the model name.

All models must use standard five emotion categories and are evaluated in standard leave-one-session-out (LOSO). See the papers for references.

Papers

Showing 2650 of 180 papers

TitleStatusHype
Dynamic Graph Neural ODE Network for Multi-modal Emotion Recognition in Conversation0
SDR-GNN: Spectral Domain Reconstruction Graph Neural Network for Incomplete Multimodal Learning in Conversational Emotion RecognitionCode1
CMATH: Cross-Modality Augmented Transformer with Hierarchical Variational Distillation for Multimodal Emotion Recognition in Conversation0
Smile upon the Face but Sadness in the Eyes: Emotion Recognition based on Facial Expressions and Eye Behaviors0
PHemoNet: A Multimodal Network for Physiological SignalsCode2
Hierarchical Hypercomplex Network for Multimodal Emotion RecognitionCode2
Early Joint Learning of Emotion Information Makes MultiModal Model Understand You Better0
Recent Trends of Multimodal Affective Computing: A Survey from NLP PerspectiveCode2
Multimodal Emotion Recognition with Vision-language Prompting and Modality Dropout0
Leveraging Contrastive Learning and Self-Training for Multimodal Emotion Recognition with Limited Labeled SamplesCode0
Video Emotion Open-vocabulary Recognition Based on Multimodal Large Language Model0
SZTU-CMU at MER2024: Improving Emotion-LLaMA with Conv-Attention for Multimodal Emotion RecognitionCode4
Enhancing Modal Fusion by Alignment and Label Matching for Multimodal Emotion RecognitionCode1
Multi Teacher Privileged Knowledge Distillation for Multimodal Expression RecognitionCode0
Tracing Intricate Cues in Dialogue: Joint Graph Structure and Sentiment Dynamics for Multimodal Emotion RecognitionCode1
Multimodal Emotion Recognition using Audio-Video Transformer Fusion with Cross AttentionCode1
MicroEmo: Time-Sensitive Multimodal Emotion Recognition with Micro-Expression Dynamics in Video Dialogues0
Masked Graph Learning with Recurrent Alignment for Multimodal Emotion Recognition in Conversation0
Textualized and Feature-based Models for Compound Multimodal Emotion Recognition in the WildCode0
Enhancing Emotion Recognition in Incomplete Data: A Novel Cross-Modal Alignment, Reconstruction, and Refinement Framework0
Emotion-LLaMA: Multimodal Emotion Recognition and Reasoning with Instruction TuningCode4
Multimodal Emotion Recognition based on Facial Expressions, Speech, and EEG0
Inconsistency-Aware Cross-Attention for Audio-Visual Fusion in Dimensional Emotion Recognition0
Empathy Through Multimodality in Conversational Interfaces0
MultiMAE-DER: Multimodal Masked Autoencoder for Dynamic Emotion RecognitionCode1
Show:102550
← PrevPage 2 of 8Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1GraphSmileWeighted F186.52Unverified
2JoyfulWeighted F185.7Unverified
3COGMENWeighted F184.5Unverified
4DANNAccuracy82.7Unverified
5MMERAccuracy81.7Unverified
6PATHOSnet v2Accuracy80.4Unverified
7Self-attention weight correction (A+T)Accuracy76.8Unverified
8CHFusionAccuracy76.5Unverified
9bc-LSTMWeighted F174.1Unverified
10Audio + Text (Stage III)F170.5Unverified
#ModelMetricClaimedVerifiedStatus
1GraphSmileWeighted F166.71Unverified
2Audio + Text (Stage III)Weighted F165.8Unverified
3JoyfulWeighted F161.77Unverified
#ModelMetricClaimedVerifiedStatus
1GraphSmileWeighted F172.81Unverified
2JoyfulWeighted F170.5Unverified
#ModelMetricClaimedVerifiedStatus
1GraphSmileWeighted F144.93Unverified
#ModelMetricClaimedVerifiedStatus
1GraphSmileWeighted F166.73Unverified
#ModelMetricClaimedVerifiedStatus
1SMPLify-Xv2v error52.9Unverified
#ModelMetricClaimedVerifiedStatus
1GraphSmileWeighted F174.31Unverified