SOTAVerified

Multimodal Emotion Recognition

This is a leaderboard for multimodal emotion recognition on the IEMOCAP dataset. The modality abbreviations are A: Acoustic T: Text V: Visual

Please include the modality in the bracket after the model name.

All models must use standard five emotion categories and are evaluated in standard leave-one-session-out (LOSO). See the papers for references.

Papers

Showing 51100 of 180 papers

TitleStatusHype
FV2ES: A Fully End2End Multimodal System for Fast Yet Effective Video Emotion Recognition InferenceCode1
GA2MIF: Graph and Attention Based Two-Stage Multi-Source Information Fusion for Conversational Emotion DetectionCode1
DialogueRNN: An Attentive RNN for Emotion Detection in ConversationsCode1
GPT-4V with Emotion: A Zero-shot Benchmark for Generalized Emotion RecognitionCode1
GraphCFC: A Directed Graph Based Cross-Modal Feature Complementation Approach for Multimodal Conversational Emotion RecognitionCode1
Group Gated Fusion on Attention-based Bidirectional Alignment for Multimodal Emotion RecognitionCode1
Knowledge-aware Bayesian Co-attention for Multimodal Emotion Recognition0
Leveraging Label Information for Multimodal Emotion Recognition0
Leveraging Label Potential for Enhanced Multimodal Emotion Recognition0
LLM supervised Pre-training for Multimodal Emotion Recognition in Conversations0
LMR-CBT: Learning Modality-fused Representations with CB-Transformer for Multimodal Emotion Recognition from Unaligned Multimodal Sequences0
M3ER: Multiplicative Multimodal Emotion Recognition Using Facial, Textual, and Speech Cues0
MART: Masked Affective RepresenTation Learning via Masked Temporal Distribution Distillation0
Masked Graph Learning with Recurrent Alignment for Multimodal Emotion Recognition in Conversation0
MEmoBERT: Pre-training Model with Prompt-based Learning for Multimodal Emotion Recognition0
MicroEmo: Time-Sensitive Multimodal Emotion Recognition with Micro-Expression Dynamics in Video Dialogues0
Modality Influence in Multimodal Machine Learning0
Multilevel Transformer For Multimodal Emotion Recognition0
Video Emotion Open-vocabulary Recognition Based on Multimodal Large Language Model0
Accommodating Missing Modalities in Time-Continuous Multimodal Emotion Recognition0
A Comparison of Time-based Models for Multimodal Emotion Recognition0
A Contextualized Real-Time Multimodal Emotion Recognition for Conversational Agents using Graph Convolutional Networks in Reinforcement Learning0
Adversarial Representation with Intra-Modal and Inter-Modal Graph Contrastive Learning for Multimodal Emotion Recognition0
A Multibias-mitigated and Sentiment Knowledge Enriched Transformer for Debiasing in Multimodal Conversational Emotion Recognition0
A Multimodal Emotion Recognition System: Integrating Facial Expressions, Body Movement, Speech, and Spoken Language0
Analyzing the Influence of Dataset Composition for Emotion Recognition0
An Audio-Video Deep and Transfer Learning Framework for Multimodal Emotion Recognition in the wild0
An Empirical Study and Improvement for Speech Emotion Recognition0
A Novel Approach to for Multimodal Emotion Recognition : Multimodal semantic information fusion0
A Robust Incomplete Multimodal Low-Rank Adaptation Approach for Emotion Recognition0
A Two-Stage Multimodal Emotion Recognition Model Based on Graph Contrastive Learning0
A Unified Transformer-based Network for multimodal Emotion Recognition0
BeMERC: Behavior-Aware MLLM-based Framework for Multimodal Emotion Recognition in Conversation0
Bias and Fairness on Multimodal Emotion Detection Algorithms0
CMATH: Cross-Modality Augmented Transformer with Hierarchical Variational Distillation for Multimodal Emotion Recognition in Conversation0
COLD Fusion: Calibrated and Ordinal Latent Distribution Fusion for Uncertainty-Aware Multimodal Emotion Recognition0
Context-aware Cascade Attention-based RNN for Video Emotion Recognition0
Context-Dependent Domain Adversarial Neural Network for Multimodal Emotion Recognition0
Contextual Dependencies in Time-Continuous Multidimensional Affect Recognition0
Continuous Multimodal Emotion Recognition Approach for AVEC 20170
Continuous-Time Audiovisual Fusion with Recurrence vs. Attention for In-The-Wild Affect Recognition0
Convolutional Attention Networks for Multimodal Emotion Recognition from Speech and Text Data0
Cross-modal Context Fusion and Adaptive Graph Convolutional Network for Multimodal Conversational Emotion Recognition0
cross-modal fusion techniques for utterance-level emotion recognition from text and speech0
CSAT‑FTCN: A Fuzzy‑Oriented Model with Contextual Self‑attention Network for Multimodal Emotion Recognition0
Deep CNN with late fusion for realtime multimodal emotion recognition0
Deep Imbalanced Learning for Multimodal Emotion Recognition in Conversations0
DER-GCN: Dialogue and Event Relation-Aware Graph Convolutional Neural Network for Multimodal Dialogue Emotion Recognition0
Do Multimodal Emotion Recognition Models Tackle Ambiguity?0
Dynamic Graph Neural ODE Network for Multi-modal Emotion Recognition in Conversation0
Show:102550
← PrevPage 2 of 4Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1GraphSmileWeighted F186.52Unverified
2JoyfulWeighted F185.7Unverified
3COGMENWeighted F184.5Unverified
4DANNAccuracy82.7Unverified
5MMERAccuracy81.7Unverified
6PATHOSnet v2Accuracy80.4Unverified
7Self-attention weight correction (A+T)Accuracy76.8Unverified
8CHFusionAccuracy76.5Unverified
9bc-LSTMWeighted F174.1Unverified
10Audio + Text (Stage III)F170.5Unverified
#ModelMetricClaimedVerifiedStatus
1GraphSmileWeighted F166.71Unverified
2Audio + Text (Stage III)Weighted F165.8Unverified
3JoyfulWeighted F161.77Unverified
#ModelMetricClaimedVerifiedStatus
1GraphSmileWeighted F172.81Unverified
2JoyfulWeighted F170.5Unverified
#ModelMetricClaimedVerifiedStatus
1GraphSmileWeighted F144.93Unverified
#ModelMetricClaimedVerifiedStatus
1GraphSmileWeighted F166.73Unverified
#ModelMetricClaimedVerifiedStatus
1SMPLify-Xv2v error52.9Unverified
#ModelMetricClaimedVerifiedStatus
1GraphSmileWeighted F174.31Unverified