Spatiotemporal Networks for Video Emotion Recognition

04/03/2017
by   Lijie Fan, et al.
0

Our experiment adapts several popular deep learning methods as well as some traditional methods on the problem of video emotion recognition. In our experiment, we use the CNN-LSTM architecture for visual information extraction and classification and utilize traditional methods such as for audio feature classification. For multimodal fusion, we use the traditional Support Vector Machine. Our experiment yields a good result on the AFEW 6.0 Dataset.

READ FULL TEXT
POST COMMENT

Comments

There are no comments yet.

Authors

09/13/2018

Investigation of Multimodal Features, Classifiers and Fusion Methods for Emotion Recognition

Automatic emotion recognition is a challenging task. In this paper, we p...
11/21/2018

Towards Emotion Recognition: A Persistent Entropy Application

Emotion recognition and classification is a very active area of research...
01/16/2017

Auxiliary Multimodal LSTM for Audio-visual Speech Recognition and Lipreading

The Aduio-visual Speech Recognition (AVSR) which employs both the video ...
03/28/2016

Audio Visual Emotion Recognition with Temporal Alignment and Perception Attention

This paper focuses on two key problems for audio-visual emotion recognit...
10/30/2018

Deep Learning as Feature Encoding for Emotion Recognition

Deep learning is popular as an end-to-end framework extracting the promi...
02/02/2022

Interpretability for Multimodal Emotion Recognition using Concept Activation Vectors

Multimodal Emotion Recognition refers to the classification of input vid...
09/09/2016

Distributed Processing of Biosignal-Database for Emotion Recognition with Mahout

This paper investigates the use of distributed processing on the problem...
This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.