Classification of Emotions and Evaluation of Customer Satisfaction from Speech in Real World Acoustic Environments

This paper focuses on finding suitable features to robustly recognize emotions and evaluate customer satisfaction from speech in real acoustic scenarios. The classification of emotions is based on standard and well-known corpora and the evaluation of customer satisfaction is based on recordings of real opinions given by customers about the received service during phone calls with call-center agents. The feature sets considered in this study include two speaker models, namely x-vectors and i-vectors, and also the well known feature set introduced in the Interspeech 2010 Paralinguistics Challenge (I2010PC). Additionally, we introduce the use of phonation, articulation and prosody features extracted with the DisVoice framework as alternative feature sets to robustly model emotions and customer satisfaction from speech. The results indicate that the I2010PC feature set is the best approach to classify emotions in the standard databases typically used in the literature. When considering the recordings collected in the call-center, without any control over the acoustic conditions, the best results are obtained with our articulation features. The I2010PC feature set includes 1584 measures while the articulation approach only includes 488 measures. We think that the proposed approach is more suitable for real-world applications where the acoustic conditions are not controlled and also it is potentially more convenient for industrial applications.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/31/2021

Graph Neural Networks to Predict Customer Satisfaction Following Interactions with a Corporate Call Center

Customer satisfaction is an important factor in creating and maintaining...
research
02/14/2023

Anchorage: Visual Analysis of Satisfaction in Customer Service Videos via Anchor Events

Delivering customer services through video communications has brought ne...
research
09/30/2020

Embedded Emotions – A Data Driven Approach to Learn Transferable Feature Representations from Raw Speech Input for Emotion Recognition

Traditional approaches to automatic emotion recognition are relying on t...
research
05/25/2020

Speaker and Posture Classification using Instantaneous Intraspeech Breathing Features

Acoustic features extracted from speech are widely used in problems such...
research
07/21/2020

AQuA: Automatic Quality Analysis of Conversational Scripts in Real-time

Call center is a point of contact between customers and the organization...
research
08/06/2021

Tell me a story about yourself: The words of shopping experience and self-satisfaction

In this paper we investigate the verbal expression of shopping experienc...
research
05/23/2023

CALLS: Japanese Empathetic Dialogue Speech Corpus of Complaint Handling and Attentive Listening in Customer Center

We present CALLS, a Japanese speech corpus that considers phone calls in...

Please sign up or login with your details

Forgot password? Click here to reset