SOTAVerified

Emotion Recognition

Emotion Recognition is an important area of research to enable effective human-computer interaction. Human emotions can be detected using speech signal, facial expressions, body language, and electroencephalography (EEG). Source: Using Deep Autoencoders for Facial Expression Recognition

Papers

Showing 17511800 of 2041 papers

TitleStatusHype
Multimodal Speech Emotion Recognition Using Audio and TextCode0
DepecheMood++: a Bilingual Emotion Lexicon Built Through Simple Yet Powerful TechniquesCode0
Text-based Sentiment Analysis and Music Emotion Recognition0
MELD: A Multimodal Multi-Party Dataset for Emotion Recognition in ConversationsCode0
BrainT at IEST 2018: Fine-tuning Multiclass Perceptron For Implicit Emotion ClassificationCode0
USI-IR at IEST 2018: Sequence Modeling and Pseudo-Relevance Feedback for Implicit Emotion Detection0
NL-FIIT at IEST-2018: Emotion Recognition utilizing Neural Networks and Multi-level PreprocessingCode0
UTFPR at IEST 2018: Exploring Character-to-Word Composition for Emotion Analysis0
SINAI at IEST 2018: Neural Encoding of Emotional External Knowledge for Emotion Classification0
HUMIR at IEST-2018: Lexicon-Sensitive and Left-Right Context-Sensitive BiLSTM for Implicit Emotion Recognition0
HGSGNLP at IEST 2018: An Ensemble of Machine Learning and Deep Neural Architectures for Implicit Emotion Classification in Tweets0
Leveraging Writing Systems Change for Deep Learning Based Chinese Emotion Analysis0
Multi-Modal Sequence Fusion via Recursive Attention for Emotion Recognition0
Interpretable Emoji Prediction via Label-Wise Attention LSTMs0
ICON: Interactive Conversational Memory Network for Multimodal Emotion Detection0
Fine-Grained Emotion Detection in Health-Related Online Posts0
CARER: Contextualized Affect Representations for Emotion RecognitionCode0
Joint Learning for Emotion Classification and Emotion Cause Detection0
Entropy-Assisted Multi-Modal Emotion Recognition Framework Based on Physiological Signals0
Investigation of Multimodal Features, Classifiers and Fusion Methods for Emotion RecognitionCode0
Convolutional Neural Network Approach for EEG-based Emotion Recognition using Brain Connectivity and its Spatial Information0
Training Deep Neural Networks with Different Datasets In-the-wild: The Emotion Recognition Paradigm0
Label-less Learning for Traffic Control in an Edge Network0
EmotiW 2018: Audio-Video, Student Engagement and Group-Level Affect Prediction0
Finding Good Representations of Emotions for Text Classification0
Emotion Recognition in Speech using Cross-Modal Transfer in the Wild0
Multimodal Local-Global Ranking Fusion for Emotion Recognition0
Multimodal Language Analysis with Recurrent Multistage FusionCode0
A Survey on Sentiment and Emotion Analysis for Computational Literary Studies0
An Occam's Razor View on Learning Audiovisual Emotion Recognition with Small Training Sets0
Contemplating Visual Emotions: Understanding and Overcoming Dataset Bias0
Normalization Before Shaking Toward Learning Symmetrically Distributed Representation Without Margin in Speech Emotion Recognition0
Hybrid Attention based Multimodal Network for Spoken Language Classification0
Who Feels What and Why? Annotation of a Literature Corpus with Semantic Roles of Emotions0
CAKE: Compact and Accurate K-dimensional representation of Emotion0
Semi-supervised Deep Generative Modelling of Incomplete Multi-Modality Emotional Data0
Emotion Recognition from Speech based on Relevant Feature and Majority Voting0
An Attention Model for group-level emotion recognitionCode0
Real time P, QRS and T wave detection by QRS matched filter methodCode0
Unsupervised Counselor Dialogue Clustering for Positive Emotion Elicitation in Neural Dialogue System0
SocialNLP 2018 EmotionX Challenge Overview: Recognizing Emotions in Dialogues0
Recognizing Emotions in Video Using Multimodal DNN Feature Fusion0
EmotionX-Area66: Predicting Emotions in Dialogues using Hierarchical Attention Network with Sequence Labeling0
EmotionX-AR: CNN-DCNN autoencoder based Emotion Classifier0
Multimodal Language Analysis in the Wild: CMU-MOSEI Dataset and Interpretable Dynamic Fusion Graph0
Construction of a Chinese Corpus for the Analysis of the Emotionality of Metaphorical Expressions0
Obtaining Reliable Human Ratings of Valence, Arousal, and Dominance for 20,000 English Words0
It All Matters: Reporting Accuracy, Inference Time and Power Consumption for Face Emotion Recognition on Embedded Systems0
A Novel Geometric Framework on Gram Matrix Trajectories for Human Behavior Understanding0
Evaluating Gammatone Frequency Cepstral Coefficients with Neural Networks for Emotion Recognition from SpeechCode0
Show:102550
← PrevPage 36 of 41Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1M2D-CLAPEmoA77.4Unverified
2M2D2EmoA76.7Unverified
3M2DEmoA76.1Unverified
4Jukebox (Pre-training: CALM)EmoA72.1Unverified
5CLMR (Pre-training: contrastive)EmoA67.8Unverified
#ModelMetricClaimedVerifiedStatus
1LogisticRegression on posteriors of xlsr-Wav2Vec2.0&bi-LSTM+AttentionAccuracy86.7Unverified
2MultiMAE-DERWAR83.61Unverified
3Intermediate-Attention-FusionAccuracy81.58Unverified
4Logistic Regression on posteriors of the CNN-14&biLSTM-GuidedSTAccuracy80.08Unverified
5ERANN-0-4Accuracy74.8Unverified
#ModelMetricClaimedVerifiedStatus
1CAGETop-3 Accuracy (%)14.73Unverified
2FocusCLIPTop-3 Accuracy (%)13.73Unverified
#ModelMetricClaimedVerifiedStatus
1VGG based5-class test accuracy66.13Unverified
#ModelMetricClaimedVerifiedStatus
1MaSaC-ERC-ZF1-score (Weighted)51.17Unverified
#ModelMetricClaimedVerifiedStatus
1BiHDMAccuracy40.34Unverified
#ModelMetricClaimedVerifiedStatus
1w2v2-L-robust-12Concordance correlation coefficient (CCC)0.64Unverified
#ModelMetricClaimedVerifiedStatus
14D-aNNAccuracy96.1Unverified
#ModelMetricClaimedVerifiedStatus
1CNN1'"1Unverified