SOTAVerified

Emotion Recognition

Emotion Recognition is an important area of research to enable effective human-computer interaction. Human emotions can be detected using speech signal, facial expressions, body language, and electroencephalography (EEG). Source: Using Deep Autoencoders for Facial Expression Recognition

Papers

Showing 14011425 of 2041 papers

TitleStatusHype
FERNet: Fine-grained Extraction and Reasoning Network for Emotion Recognition in Dialogues0
MemoSYS at SemEval-2020 Task 8: Multimodal Emotion Analysis in Memes0
SemEval-2020 Task 8: Memotion Analysis- the Visuo-Lingual Metaphor!0
CN-HIT-MI.T at SemEval-2020 Task 8: Memotion Analysis Based on BERT0
Hitachi at SemEval-2020 Task 8: Simple but Effective Modality Ensemble for Meme Emotion Recognition0
Summarize before Aggregate: A Global-to-local Heterogeneous Graph Inference Network for Conversational Emotion Recognition0
An Iterative Emotion Interaction Network for Emotion Recognition in Conversations0
Towards Label-Agnostic Emotion Embeddings0
Emotional Semantics-Preserved and Feature-Aligned CycleGAN for Visual Emotion Adaptation0
Deep Learning in EEG: Advance of the Last Ten-Year Critical Period0
Self-Supervised learning with cross-modal transformers for emotion recognition0
Deep Residual Local Feature Learning for Speech Emotion Recognition0
On the use of Self-supervised Pre-trained Acoustic and Linguistic Features for Continuous Speech Emotion Recognition0
Continuous Emotion Recognition with Spatiotemporal Convolutional Neural Networks0
Interpretable Image Emotion Recognition: A Domain Adaptation Approach Using Facial Expressions0
Improving Multimodal Accuracy Through Modality Pre-training and Attention0
WaDeNet: Wavelet Decomposition based CNN for Speech Processing0
Recognizing More Emotions with Less Data Using Self-supervised Transfer Learning0
NUAA-QMUL at SemEval-2020 Task 8: Utilizing BERT and DenseNet for Internet Meme Emotion AnalysisCode0
Robust Latent Representations via Cross-Modal Translation and Alignment0
Experiencers, Stimuli, or Targets: Which Semantic Roles Enable Machine Learning to Infer the Emotions?0
I miss you babe: Analyzing Emotion Dynamics During COVID-19 Pandemic0
Identifying Worry in Twitter: Beyond Emotion Analysis0
Analysis of Resource-efficient Predictive Models for Natural Language Processing0
Relation-aware Graph Attention Networks with Relational Position Encodings for Emotion Recognition in ConversationsCode1
Show:102550
← PrevPage 57 of 82Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1M2D-CLAPEmoA77.4Unverified
2M2D2EmoA76.7Unverified
3M2DEmoA76.1Unverified
4Jukebox (Pre-training: CALM)EmoA72.1Unverified
5CLMR (Pre-training: contrastive)EmoA67.8Unverified
#ModelMetricClaimedVerifiedStatus
1LogisticRegression on posteriors of xlsr-Wav2Vec2.0&bi-LSTM+AttentionAccuracy86.7Unverified
2MultiMAE-DERWAR83.61Unverified
3Intermediate-Attention-FusionAccuracy81.58Unverified
4Logistic Regression on posteriors of the CNN-14&biLSTM-GuidedSTAccuracy80.08Unverified
5ERANN-0-4Accuracy74.8Unverified
#ModelMetricClaimedVerifiedStatus
1CAGETop-3 Accuracy (%)14.73Unverified
2FocusCLIPTop-3 Accuracy (%)13.73Unverified
#ModelMetricClaimedVerifiedStatus
1VGG based5-class test accuracy66.13Unverified
#ModelMetricClaimedVerifiedStatus
1MaSaC-ERC-ZF1-score (Weighted)51.17Unverified
#ModelMetricClaimedVerifiedStatus
1BiHDMAccuracy40.34Unverified
#ModelMetricClaimedVerifiedStatus
1w2v2-L-robust-12Concordance correlation coefficient (CCC)0.64Unverified
#ModelMetricClaimedVerifiedStatus
14D-aNNAccuracy96.1Unverified
#ModelMetricClaimedVerifiedStatus
1CNN1'"1Unverified