Video Multimodal Emotion Recognition System for Real World Applications

08/28/2023
by   Sun-Kyung Lee, et al.
0

This paper proposes a system capable of recognizing a speaker's utterance-level emotion through multimodal cues in a video. The system seamlessly integrates multiple AI models to first extract and pre-process multimodal information from the raw video input. Next, an end-to-end MER model sequentially predicts the speaker's emotions at the utterance level. Additionally, users can interactively demonstrate the system through the implemented interface.

READ FULL TEXT

page 1

page 2

research
03/05/2021

Analyzing the Influence of Dataset Composition for Emotion Recognition

Recognizing emotions from text in multimodal architectures has yielded p...
research
02/02/2022

Interpretability for Multimodal Emotion Recognition using Concept Activation Vectors

Multimodal Emotion Recognition refers to the classification of input vid...
research
06/05/2022

A Multimodal Corpus for Emotion Recognition in Sarcasm

While sentiment and emotion analysis have been studied extensively, the ...
research
05/03/2018

Transformer for Emotion Recognition

This paper describes the UMONS solution for the OMG-Emotion Challenge. W...
research
07/14/2020

DeepMSRF: A novel Deep Multimodal Speaker Recognition framework with Feature selection

For recognizing speakers in video streams, significant research studies ...
research
10/20/2019

Speech Emotion Recognition with Dual-Sequence LSTM Architecture

Speech Emotion Recognition (SER) has emerged as a critical component of ...
research
05/03/2018

Abstract: UMONS submission for the OMG-Emotion Challenge

This paper describes the UMONS solution for the OMG-Emotion Challenge. W...

Please sign up or login with your details

Forgot password? Click here to reset