It is Okay to Not Be Okay: Overcoming Emotional Bias in Affective Image Captioning by Contrastive Data Collection

04/15/2022
by   Youssef Mohamed, et al.
0

Datasets that capture the connection between vision, language, and affection are limited, causing a lack of understanding of the emotional aspect of human intelligence. As a step in this direction, the ArtEmis dataset was recently introduced as a large-scale dataset of emotional reactions to images along with language explanations of these chosen emotions. We observed a significant emotional bias towards instance-rich emotions, making trained neural speakers less accurate in describing under-represented emotions. We show that collecting new data, in the same way, is not effective in mitigating this emotional bias. To remedy this problem, we propose a contrastive data collection approach to balance ArtEmis with a new complementary dataset such that a pair of similar images have contrasting emotions (one positive and one negative). We collected 260,533 instances using the proposed method, we combine them with ArtEmis, creating a second iteration of the dataset. The new combined dataset, dubbed ArtEmis v2.0, has a balanced distribution of emotions with explanations revealing more fine details in the associated painting. Our experiments show that neural speakers trained on the new dataset improve CIDEr and METEOR evaluation metrics by 20 Finally, we also show that the performance per emotion of neural speakers is improved across all the emotion categories, significantly on under-represented emotions. The collected dataset and code are available at https://artemisdataset-v2.org.

READ FULL TEXT

page 1

page 3

page 5

page 6

page 8

research
08/30/2018

MES-P: an Emotional Tonal Speech Dataset in Mandarin Chinese with Distal and Proximal Labels

Emotion shapes all aspects of our interpersonal and intellectual experie...
research
12/07/2021

Multi-speaker Emotional Text-to-speech Synthesizer

We present a methodology to train our multi-speaker emotional text-to-sp...
research
03/12/2022

Taking an Emotional Look at Video Paragraph Captioning

Translating visual data into natural language is essential for machines ...
research
12/21/2022

Contrastive Language-Vision AI Models Pretrained on Web-Scraped Multimodal Data Exhibit Sexual Objectification Bias

Nine language-vision AI models trained on web scrapes with the Contrasti...
research
01/19/2021

ArtEmis: Affective Language for Visual Art

We present a novel large-scale dataset and accompanying machine learning...
research
05/02/2020

Are Emojis Emotional? A Study to Understand the Association between Emojis and Emotions

Given the growing ubiquity of emojis in language, there is a need for me...
research
07/10/2016

Towards an "In-the-Wild" Emotion Dataset Using a Game-based Framework

In order to create an "in-the-wild" dataset of facial emotions with larg...

Please sign up or login with your details

Forgot password? Click here to reset