Emotion Recognition From Speech With Recurrent Neural Networks

01/27/2017
by   Vladimir Chernykh, et al.
0

In this paper the task of emotion recognition from speech is considered. Proposed approach uses deep recurrent neural network trained on a sequence of acoustic features calculated over small speech intervals. At the same time special probabilistic-nature CTC loss function allows to consider long utterances containing both emotional and unemotional parts. The effectiveness of such an approach is shown in two ways. First one is the comparison with recent advances in this field. While second way implies measuring human performance on the same task, which also was done by authors.

READ FULL TEXT

page 3

page 10

research
03/30/2018

Reusing Neural Speech Representations for Auditory Emotion Recognition

Acoustic emotion recognition aims to categorize the affective state of t...
research
03/03/2020

The Effect of Silence Feature in Dimensional Speech Emotion Recognition

Silence is a part of human-to-human communication, which can be a clue f...
research
07/28/2022

Extending RNN-T-based speech recognition systems with emotion and language classification

Speech transcription, emotion recognition, and language identification a...
research
11/07/2022

Hi,KIA: A Speech Emotion Recognition Dataset for Wake-Up Words

Wake-up words (WUW) is a short sentence used to activate a speech recogn...
research
02/24/2016

How Deep Neural Networks Can Improve Emotion Recognition on Video Data

We consider the task of dimensional emotion recognition on video data us...
research
06/11/2019

Focal Loss based Residual Convolutional Neural Network for Speech Emotion Recognition

This paper proposes a Residual Convolutional Neural Network (ResNet) bas...
research
10/14/2022

Training speech emotion classifier without categorical annotations

There are two paradigms of emotion representation, categorical labeling ...

Please sign up or login with your details

Forgot password? Click here to reset