Predicting Depression Severity by Multi-Modal Feature Engineering and Fusion
2017-11-29Unverified0· sign in to hype
Aven Samareh, Yan Jin, Zhangyang Wang, Xiangyu Chang, Shuai Huang
Unverified — Be the first to reproduce this paper.
ReproduceAbstract
We present our preliminary work to determine if patient's vocal acoustic, linguistic, and facial patterns could predict clinical ratings of depression severity, namely Patient Health Questionnaire depression scale (PHQ-8). We proposed a multi modal fusion model that combines three different modalities: audio, video , and text features. By training over AVEC 2017 data set, our proposed model outperforms each single modality prediction model, and surpasses the data set baseline with ice margin.