Depression Severity Estimation from Multiple Modalities

11/10/2017
by   Evgeny Stepanov, et al.
0

Depression is a major debilitating disorder which can affect people from all ages. With a continuous increase in the number of annual cases of depression, there is a need to develop automatic techniques for the detection of the presence and extent of depression. In this AVEC challenge we explore different modalities (speech, language and visual features extracted from face) to design and develop automatic methods for the detection of depression. In psychology literature, the PHQ-8 questionnaire is well established as a tool for measuring the severity of depression. In this paper we aim to automatically predict the PHQ-8 scores from features extracted from the different modalities. We show that visual features extracted from facial landmarks obtain the best performance in terms of estimating the PHQ-8 results with a mean absolute error (MAE) of 4.66 on the development set. Behavioral characteristics from speech provide an MAE of 4.73. Language features yield a slightly higher MAE of 5.17. When switching to the test set, our Turn Features derived from audio transcriptions achieve the best performance, scoring an MAE of 4.11 (corresponding to an RMSE of 4.94), which makes our system the winner of the AVEC 2017 depression sub-challenge.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/18/2017

Continuous Multimodal Emotion Recognition Approach for AVEC 2017

This paper reports the analysis of audio and visual features in predicti...
research
09/16/2019

MFCC-based Recurrent Neural Network for Automatic Clinical Depression Recognition and Assessment from Speech

Major depression, also known as clinical depression, is a constant sense...
research
04/24/2015

Cultural Event Recognition with Visual ConvNets and Temporal Models

This paper presents our contribution to the ChaLearn Challenge 2015 on C...
research
09/30/2020

Embedded Emotions – A Data Driven Approach to Learn Transferable Feature Representations from Raw Speech Input for Emotion Recognition

Traditional approaches to automatic emotion recognition are relying on t...
research
07/21/2023

Poverty rate prediction using multi-modal survey and earth observation data

This work presents an approach for combining household demographic and l...
research
09/11/2020

Multimodal Depression Severity Prediction from medical bio-markers using Machine Learning Tools and Technologies

Depression has been a leading cause of mental-health illnesses across th...
research
10/11/2022

Aggregating Layers for Deepfake Detection

The increasing popularity of facial manipulation (Deepfakes) and synthet...

Please sign up or login with your details

Forgot password? Click here to reset