Semi-supervised Multi-modal Emotion Recognition with Cross-Modal Distribution Matching

09/05/2020
by   Jingjun Liang, et al.
0

Automatic emotion recognition is an active research topic with wide range of applications. Due to the high manual annotation cost and inevitable label ambiguity, the development of emotion recognition dataset is limited in both scale and quality. Therefore, one of the key challenges is how to build effective models with limited data resource. Previous works have explored different approaches to tackle this challenge including data enhancement, transfer learning, and semi-supervised learning etc. However, the weakness of these existing approaches includes such as training instability, large performance loss during transfer, or marginal improvement. In this work, we propose a novel semi-supervised multi-modal emotion recognition model based on cross-modality distribution matching, which leverages abundant unlabeled data to enhance the model training under the assumption that the inner emotional status is consistent at the utterance level across modalities. We conduct extensive experiments to evaluate the proposed model on two benchmark datasets, IEMOCAP and MELD. The experiment results prove that the proposed semi-supervised learning model can effectively utilize unlabeled data and combine multi-modalities to boost the emotion recognition performance, which outperforms other state-of-the-art approaches under the same condition. The proposed model also achieves competitive capacity compared with existing approaches which take advantage of additional auxiliary information such as speaker and interaction context.

READ FULL TEXT

page 3

page 8

research
04/25/2017

Semi-supervised Bayesian Deep Multi-modal Emotion Recognition

In emotion recognition, it is difficult to recognize human's emotional s...
research
11/11/2021

Multilingual and Multilabel Emotion Recognition using Virtual Adversarial Training

Virtual Adversarial Training (VAT) has been effective in learning robust...
research
04/18/2023

MER 2023: Multi-label Learning, Modality Robustness, and Semi-Supervised Learning

Over the past few decades, multimodal emotion recognition has made remar...
research
09/19/2022

Multi-Task Vision Transformer for Semi-Supervised Driver Distraction Detection

Driver distraction detection is an important computer vision problem tha...
research
07/14/2020

TCGM: An Information-Theoretic Framework for Semi-Supervised Multi-Modality Learning

Fusing data from multiple modalities provides more information to train ...
research
07/27/2018

Semi-supervised Deep Generative Modelling of Incomplete Multi-Modality Emotional Data

There are threefold challenges in emotion recognition. First, it is diff...
research
07/13/2019

Multi-Task Semi-Supervised Adversarial Autoencoding for Speech Emotion

Despite the widespread use of supervised deep learning methods for affec...

Please sign up or login with your details

Forgot password? Click here to reset