Contribution of Glottal Waveform in Speech Emotion: A Comparative Pairwise Investigation

08/30/2018
by   Zhongzhe Xiao, et al.
0

In this work, we investigated the contribution of the glottal waveform in human vocal emotion expressing. Seven emotional states including moderate and intense versions of three emotional families as anger, joy, and sadness, plus a neutral state are considered, with speech samples in Mandarin Chinese. The glottal waveform extracted from speech samples of different emotion states are first analyzed in both time domain and frequency domain to discover their differences. Comparative emotion classifications are then taken out based on features extracted from original whole speech signal and only glottal wave signal. In experiments of generation of a performance-driven hierarchical classifier architecture, and pairwise classification on individual emotional states, the low difference between accuracies obtained from speech signal and glottal signal proved that a majority of emotional cues in speech could be conveyed through glottal waveform. The best distinguishable emotional pair by glottal waveform is intense anger against moderate sadness, with the accuracy of 92.45 better valence cues than arousal cues of emotion.

READ FULL TEXT
research
09/03/2019

Multiresolution analysis (discrete wavelet transform) through Daubechies family for emotion recognition in speech

We propose a study of the mathematical properties of voice as an audio s...
research
08/30/2018

MES-P: an Emotional Tonal Speech Dataset in Mandarin Chinese with Distal and Proximal Labels

Emotion shapes all aspects of our interpersonal and intellectual experie...
research
08/12/2020

Emotion Profile Refinery for Speech Emotion Classification

Human emotions are inherently ambiguous and impure. When designing syste...
research
03/31/2018

Speaker Verification in Emotional Talking Environments based on Three-Stage Framework

This work is dedicated to introducing, executing, and assessing a three-...
research
09/27/2021

Emotional Speech Synthesis for Companion Robot to Imitate Professional Caregiver Speech

When people try to influence others to do something, they subconsciously...
research
06/04/2019

ShEMO -- A Large-Scale Validated Database for Persian Speech Emotion Detection

This paper introduces a large-scale, validated database for Persian call...
research
10/11/2021

LaughNet: synthesizing laughter utterances from waveform silhouettes and a single laughter example

Emotional and controllable speech synthesis is a topic that has received...

Please sign up or login with your details

Forgot password? Click here to reset