SOTAVerified

Multimodal Emotion Recognition

This is a leaderboard for multimodal emotion recognition on the IEMOCAP dataset. The modality abbreviations are A: Acoustic T: Text V: Visual

Please include the modality in the bracket after the model name.

All models must use standard five emotion categories and are evaluated in standard leave-one-session-out (LOSO). See the papers for references.

Papers

Showing 101150 of 180 papers

TitleStatusHype
Dynamic Modality and View Selection for Multimodal Emotion Recognition with Missing Modalities0
Early Joint Learning of Emotion Information Makes MultiModal Model Understand You Better0
EMERSK -- Explainable Multimodal Emotion Recognition with Situational Knowledge0
EMOE: Modality-Specific Enhanced Dynamic Emotion Experts0
EmoTech: A Multi-modal Speech Emotion Recognition Using Multi-source Low-level Information with Hybrid Recurrent Network0
EmotiCon: Context-Aware Multimodal Emotion Recognition using Frege's Principle0
Emotion recognition based on multi-modal electrophysiology multi-head attention Contrastive Learning0
Emotion recognition by fusing time synchronous and time asynchronous representations0
Emotion Recognition with Pre-Trained Transformers Using Multimodal Signals0
Empathy Through Multimodality in Conversational Interfaces0
Enhancing Emotion Recognition in Incomplete Data: A Novel Cross-Modal Alignment, Reconstruction, and Refinement Framework0
Enhancing Multimodal Emotion Recognition through Multi-Granularity Cross-Modal Alignment0
Exploring Attention Mechanisms for Multimodal Emotion Recognition in an Emergency Call Center Corpus0
FAF: A novel multimodal emotion recognition approach integrating face, body and text0
Framewise approach in multimodal emotion recognition in OMG challenge0
GatedxLSTM: A Multimodal Affective Computing Approach for Emotion Recognition in Conversations0
GSDNet: Revisiting Incomplete Multimodal-Diffusion from Graph Spectrum Perspective for Conversation Emotion Recognition0
United we stand, Divided we fall: Handling Weak Complementary Relationships for Audio-Visual Emotion Recognition in Valence-Arousal Space0
HCAM -- Hierarchical Cross Attention Model for Multi-modal Emotion Recognition0
Hierarchical Audio-Visual Information Fusion with Multi-label Joint Decoding for MER 20230
Inconsistency-Aware Cross-Attention for Audio-Visual Fusion in Dimensional Emotion Recognition0
Interpretability for Multimodal Emotion Recognition using Concept Activation Vectors0
Interpretable Multimodal Emotion Recognition using Facial Features and Physiological Signals0
Investigating EEG-Based Functional Connectivity Patterns for Multimodal Emotion Recognition0
MULTI-MODAL EMOTION RECOGNITION ON IEMOCAP WITH NEURAL NETWORKS.0
Multimodal Emotion Recognition on RAVDESS Dataset Using Transfer Learning0
Multimodal Emotion Recognition Using Multimodal Deep Learning0
Multimodal Emotion Recognition using Transfer Learning from Speaker Recognition and BERT-based models0
Multimodal Emotion Recognition with Vision-language Prompting and Modality Dropout0
Multimodal End-to-End Group Emotion Recognition using Cross-Modal Attention0
Multimodal Mixture of Low-Rank Experts for Sentiment Analysis and Emotion Recognition0
MVP: Multimodal Emotion Recognition based on Video and Physiological Signals0
Noise-Resistant Multimodal Transformer for Emotion Recognition0
Progressive Modality Reinforcement for Human Multimodal Emotion Recognition From Unaligned Multimodal Sequences0
PsyCounAssist: A Full-Cycle AI-Powered Psychological Counseling Assistant System0
Revisiting Disentanglement and Fusion on Modality and Context in Conversational Multimodal Emotion Recognition0
Revisiting Multimodal Emotion Recognition in Conversation from the Perspective of Graph Spectrum0
Seamless Multimodal Biometrics for Continuous Personalised Wellbeing Monitoring0
Smile upon the Face but Sadness in the Eyes: Emotion Recognition based on Facial Expressions and Eye Behaviors0
Speech Emotion Recognition Based on Self-Attention Weight Correction for Acoustic and Text Features0
TACOformer:Token-channel compounded Cross Attention for Multimodal Emotion Recognition0
Towards Multimodal Emotion Recognition in German Speech Events in Cars using Transfer Learning0
UniMEEC: Towards Unified Multimodal Emotion Recognition and Emotion Cause0
Unimodal-driven Distillation in Multimodal Emotion Recognition with Dynamic Fusion0
Using Auxiliary Tasks In Multimodal Fusion Of Wav2vec 2.0 And BERT For Multimodal Emotion Recognition0
Using Large Pre-Trained Models with Cross-Modal Attention for Multi-Modal Emotion Recognition0
Versatile audio-visual learning for emotion recognition0
0/1 Deep Neural Networks via Block Coordinate Descent0
Multimodal Emotion-Cause Pair Extraction in Conversations0
Multimodal Emotion Recognition among Couples from Lab Settings to Daily Life using Smartwatches0
Show:102550
← PrevPage 3 of 4Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1GraphSmileWeighted F186.52Unverified
2JoyfulWeighted F185.7Unverified
3COGMENWeighted F184.5Unverified
4DANNAccuracy82.7Unverified
5MMERAccuracy81.7Unverified
6PATHOSnet v2Accuracy80.4Unverified
7Self-attention weight correction (A+T)Accuracy76.8Unverified
8CHFusionAccuracy76.5Unverified
9bc-LSTMWeighted F174.1Unverified
10Audio + Text (Stage III)F170.5Unverified
#ModelMetricClaimedVerifiedStatus
1GraphSmileWeighted F166.71Unverified
2Audio + Text (Stage III)Weighted F165.8Unverified
3JoyfulWeighted F161.77Unverified
#ModelMetricClaimedVerifiedStatus
1GraphSmileWeighted F172.81Unverified
2JoyfulWeighted F170.5Unverified
#ModelMetricClaimedVerifiedStatus
1GraphSmileWeighted F144.93Unverified
#ModelMetricClaimedVerifiedStatus
1GraphSmileWeighted F166.73Unverified
#ModelMetricClaimedVerifiedStatus
1SMPLify-Xv2v error52.9Unverified
#ModelMetricClaimedVerifiedStatus
1GraphSmileWeighted F174.31Unverified