SOTAVerified

Emotion Recognition in Conversation

Given the transcript of a conversation along with speaker information of each constituent utterance, the ERC task aims to identify the emotion of each utterance from several pre-defined emotions. Formally, given the input sequence of N number of utterances [(u1, p1), (u2, p2), . . . , (uN , pN )], where each utterance ui = [ui,1, ui,2, . . . , ui,T ] consists of T words ui,j and spoken by party pi, the task is to predict the emotion label ei of each utterance ui. .

Papers

Showing 125 of 141 papers

TitleStatusHype
Qwen-Audio: Advancing Universal Audio Understanding via Unified Large-Scale Audio-Language ModelsCode3
BERT: Pre-training of Deep Bidirectional Transformers for Language UnderstandingCode3
Recent Trends of Multimodal Affective Computing: A Survey from NLP PerspectiveCode2
UniMSE: Towards Unified Multimodal Sentiment Analysis and Emotion RecognitionCode2
CPED: A Large-Scale Chinese Personalized and Emotional Dialogue Dataset for Conversational AICode2
Structure-Aware Transformer for Graph Representation LearningCode2
COSMIC: COmmonSense knowledge for eMotion Identification in ConversationsCode2
Beyond Silent Letters: Amplifying LLMs in Emotion Recognition with Vocal NuancesCode1
Tracing Intricate Cues in Dialogue: Joint Graph Structure and Sentiment Dynamics for Multimodal Emotion RecognitionCode1
BiosERC: Integrating Biography Speakers Supported by LLMs for ERC TasksCode1
Multi-Task Multi-Modal Self-Supervised Learning for Facial Expression RecognitionCode1
Emotion-Anchored Contrastive Learning Framework for Emotion Recognition in ConversationCode1
Curriculum Learning Meets Directed Acyclic Graph for Multimodal Emotion RecognitionCode1
TelME: Teacher-leading Multimodal Fusion Network for Emotion Recognition in ConversationCode1
Joyful: Joint Modality Fusion and Graph Contrastive Learning for Multimodal Emotion RecognitionCode1
A Transformer-Based Model With Self-Distillation for Multimodal Emotion Recognition in ConversationsCode1
InstructERC: Reforming Emotion Recognition in Conversation with Multi-task Retrieval-Augmented Large Language ModelsCode1
UniSA: Unified Generative Framework for Sentiment AnalysisCode1
CFN-ESA: A Cross-Modal Fusion Network with Emotion-Shift Awareness for Dialogue Emotion RecognitionCode1
A Facial Expression-Aware Multimodal Multi-task Learning Framework for Emotion Recognition in Multi-party ConversationsCode1
Supervised Adversarial Contrastive Learning for Emotion Recognition in ConversationsCode1
How to Enhance Causal Discrimination of Utterances: A Case on Affective ReasoningCode1
Multivariate, Multi-Frequency and Multimodal: Rethinking Graph Neural Networks for Emotion Recognition in ConversationCode1
Distribution-based Emotion Recognition in ConversationCode1
Supervised Prototypical Contrastive Learning for Emotion Recognition in ConversationCode1
Show:102550
← PrevPage 1 of 6Next →

No leaderboard results yet.