SOTAVerified

MMDAG: Multimodal Directed Acyclic Graph Network for Emotion Recognition in Conversation

2022-06-01LREC 2022Unverified0· sign in to hype

Shuo Xu, Yuxiang Jia, Changyong Niu, Hongying Zan

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

Emotion recognition in conversation is important for an empathetic dialogue system to understand the user’s emotion and then generate appropriate emotional responses. However, most previous researches focus on modeling conversational contexts primarily based on the textual modality or simply utilizing multimodal information through feature concatenation. In order to exploit multimodal information and contextual information more effectively, we propose a multimodal directed acyclic graph (MMDAG) network by injecting information flows inside modality and across modalities into the DAG architecture. Experiments on IEMOCAP and MELD show that our model outperforms other state-of-the-art models. Comparative studies validate the effectiveness of the proposed modality fusion method.

Tasks

Reproductions