Predicting Depression Severity by Multi-Modal Feature Engineering and Fusion

11/29/2017
by   Aven Samareh, et al.
1

We present our preliminary work to determine if patient's vocal acoustic, linguistic, and facial patterns could predict clinical ratings of depression severity, namely Patient Health Questionnaire depression scale (PHQ-8). We proposed a multi modal fusion model that combines three different modalities: audio, video , and text features. By training over AVEC 2017 data set, our proposed model outperforms each single modality prediction model, and surpasses the data set baseline with ice margin.

READ FULL TEXT

page 1

page 2

research
09/01/2019

Video Affective Effects Prediction with Multi-modal Fusion and Shot-Long Temporal Context

Predicting the emotional impact of videos using machine learning is a ch...
research
05/29/2020

Large Scale Audiovisual Learning of Sounds with Weakly Labeled Data

Recognizing sounds is a key aspect of computational audio scene analysis...
research
03/30/2022

Speech and the n-Back task as a lens into depression. How combining both may allow us to isolate different core symptoms of depression

Embedded in any speech signal is a rich combination of cognitive, neurom...
research
04/08/2019

Text-based Depression Detection: What Triggers An Alert

Recent advances in automatic depression detection mostly derive from mod...
research
11/21/2018

Measuring Depression Symptom Severity from Spoken Language and 3D Facial Expressions

With more than 300 million people depressed worldwide, depression is a g...
research
07/13/2022

Multi-modal Depression Estimation based on Sub-attentional Fusion

Failure to timely diagnose and effectively treat depression leads to ove...
research
04/10/2022

Multimodal Machine Learning in Precision Health

As machine learning and artificial intelligence are more frequently bein...

Please sign up or login with your details

Forgot password? Click here to reset