SOTAVerified

Emotion Recognition

Emotion Recognition is an important area of research to enable effective human-computer interaction. Human emotions can be detected using speech signal, facial expressions, body language, and electroencephalography (EEG). Source: Using Deep Autoencoders for Facial Expression Recognition

Papers

Showing 17511800 of 2041 papers

TitleStatusHype
Emotion Recognition From Speech With Recurrent Neural NetworksCode0
A Change of Heart: Improving Speech Emotion Recognition through Speech-to-Text Modality ConversionCode0
Data Augmentation for Emotion Detection in Small Imbalanced Text DataCode0
VISTANet: VIsual Spoken Textual Additive Net for Interpretable Multimodal Emotion RecognitionCode0
Emotion Recognition from SpeechCode0
Multi-Task Learning Framework for Emotion Recognition in-the-wildCode0
The Skipped Beat: A Study of Sociopragmatic Understanding in LLMs for 64 LanguagesCode0
Multi-attention Recurrent Network for Human Communication ComprehensionCode0
Pretext Tasks selection for multitask self-supervised speech representation learningCode0
Hide-and-Seek: A Data Augmentation Technique for Weakly-Supervised Localization and BeyondCode0
Pretrained audio neural networks for Speech emotion recognition in PortugueseCode0
DAGER: Deep Age, Gender and Emotion Recognition Using Convolutional Neural NetworkCode0
Student Engagement Detection Using Emotion Analysis, Eye Tracking and Head Movement with Machine LearningCode0
Pre-trained Speech Processing Models Contain Human-Like Biases that Propagate to Speech Emotion RecognitionCode0
HHHFL: Hierarchical Heterogeneous Horizontal Federated Learning for ElectroencephalographyCode0
Customising General Large Language Models for Specialised Emotion Recognition TasksCode0
YNU-HPCC at SemEval-2020 Task 8: Using a Parallel-Channel Model for Memotion AnalysisCode0
Multi-label Co-regularization for Semi-supervised Facial Action Unit RecognitionCode0
BSC-UPC at EmoSPeech-IberLEF2024: Attention Pooling for Emotion RecognitionCode0
Multi-level Fusion of Wav2vec 2.0 and BERT for Multimodal Emotion RecognitionCode0
Words Can Shift: Dynamically Adjusting Word Representations Using Nonverbal BehaviorsCode0
Bridging Modalities: Knowledge Distillation and Masked Training for Translating Multi-Modal Emotion Recognition to Uni-Modal, Speech-Only Emotion RecognitionCode0
Cultural-Aware AI Model for Emotion RecognitionCode0
Using Knowledge-Embedded Attention to Augment Pre-trained Language Models for Fine-Grained Emotion RecognitionCode0
EmotionIC: emotional inertia and contagion-driven dependency modeling for emotion recognition in conversationCode0
Heterogeneity over Homogeneity: Investigating Multilingual Speech Pre-Trained Models for Detecting Audio DeepfakeCode0
Emotion Detection From Tweets Using a BERT and SVM Ensemble ModelCode0
SGED: A Benchmark dataset for Performance Evaluation of Spiking Gesture Emotion RecognitionCode0
What About Emotions? Guiding Fine-Grained Emotion Extraction from Mobile App ReviewsCode0
Bridge then Begin Anew: Generating Target-relevant Intermediate Model for Source-free Visual Emotion AdaptationCode0
Are you sure? Analysing Uncertainty Quantification Approaches for Real-world Speech Emotion RecognitionCode0
Progressive Neural Networks for Transfer Learning in Emotion RecognitionCode0
Multimodal Embeddings from Language ModelsCode0
BrainT at IEST 2018: Fine-tuning Multiclass Perceptron For Implicit Emotion ClassificationCode0
The Whole Is Bigger Than the Sum of Its Parts: Modeling Individual Annotators to Capture Emotional VariabilityCode0
Improvement in Facial Emotion Recognition using Synthetic Data Generated by Diffusion ModelCode0
A novel policy for pre-trained Deep Reinforcement Learning for Speech Emotion RecognitionCode0
CTL-MTNet: A Novel CapsNet and Transfer Learning-Based Mixed Task Net for the Single-Corpus and Cross-Corpus Speech Emotion RecognitionCode0
Crossmodal ASR Error Correction with Discrete Speech UnitsCode0
Cross Lingual Speech Emotion Recognition: Urdu vs. Western LanguagesCode0
AIRCADE: an Anechoic and IR Convolution-based Auralization Data-compilation EnsembleCode0
An Interaction-aware Attention Network for Speech Emotion Recognition in Spoken DialogsCode0
Affective-NLI: Towards Accurate and Interpretable Personality Recognition in ConversationCode0
Multi-Modal Emotion recognition on IEMOCAP Dataset using Deep LearningCode0
Cross-Lingual Speech Emotion Recognition: Humans vs. Self-Supervised ModelsCode0
An Improved StarGAN for Emotional Voice Conversion: Enhancing Voice Quality and Data AugmentationCode0
Angry Men, Sad Women: Large Language Models Reflect Gendered Stereotypes in Emotion AttributionCode0
Emotion Analysis of Writers and Readers of Japanese Tweets on VaccinationsCode0
Multimodal Emotion Recognition Using Deep Canonical Correlation AnalysisCode0
Group-level Emotion Recognition using Transfer Learning from Face IdentificationCode0
Show:102550
← PrevPage 36 of 41Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1M2D-CLAPEmoA77.4Unverified
2M2D2EmoA76.7Unverified
3M2DEmoA76.1Unverified
4Jukebox (Pre-training: CALM)EmoA72.1Unverified
5CLMR (Pre-training: contrastive)EmoA67.8Unverified
#ModelMetricClaimedVerifiedStatus
1LogisticRegression on posteriors of xlsr-Wav2Vec2.0&bi-LSTM+AttentionAccuracy86.7Unverified
2MultiMAE-DERWAR83.61Unverified
3Intermediate-Attention-FusionAccuracy81.58Unverified
4Logistic Regression on posteriors of the CNN-14&biLSTM-GuidedSTAccuracy80.08Unverified
5ERANN-0-4Accuracy74.8Unverified
#ModelMetricClaimedVerifiedStatus
1CAGETop-3 Accuracy (%)14.73Unverified
2FocusCLIPTop-3 Accuracy (%)13.73Unverified
#ModelMetricClaimedVerifiedStatus
1VGG based5-class test accuracy66.13Unverified
#ModelMetricClaimedVerifiedStatus
1MaSaC-ERC-ZF1-score (Weighted)51.17Unverified
#ModelMetricClaimedVerifiedStatus
1BiHDMAccuracy40.34Unverified
#ModelMetricClaimedVerifiedStatus
1w2v2-L-robust-12Concordance correlation coefficient (CCC)0.64Unverified
#ModelMetricClaimedVerifiedStatus
14D-aNNAccuracy96.1Unverified
#ModelMetricClaimedVerifiedStatus
1CNN1'"1Unverified