Women also Snowboard: Overcoming Bias in Captioning Models

03/26/2018
by   Kaylee Burns, et al.
0

Most machine learning methods are known to capture and exploit biases of the training data. While some biases are beneficial for learning, others are harmful. Specifically, image captioning models tend to exaggerate biases present in training data (e.g., if a word is present in 60 sentences, it might be predicted in 70 lead to incorrect captions in domains where unbiased captions are desired, or required, due to over-reliance on the learned prior and image context. In this work we investigate generation of gender-specific caption words (e.g. man, woman) based on the person's appearance or the image context. We introduce a new Equalizer model that ensures equal gender probability when gender evidence is occluded in a scene and confident predictions when gender evidence is present. The resulting model is forced to look at a person rather than use contextual cues to make a gender-specific predictions. The losses that comprise our model, the Appearance Confusion Loss and the Confident Loss, are general, and can be added to any description model in order to mitigate impacts of unwanted bias in a description dataset. Our proposed model has lower error than prior work when describing images with people and mentioning their gender and more closely matches the ground truth ratio of sentences including women to sentences including men. We also show that unlike other approaches, our model is indeed more often looking at people when predicting their gender.

READ FULL TEXT

page 2

page 5

page 15

page 19

page 20

research
07/02/2018

Women also Snowboard: Overcoming Bias in Captioning Models (Extended Abstract)

Most machine learning methods are known to capture and exploit biases of...
research
04/07/2023

Model-Agnostic Gender Debiased Image Captioning

Image captioning models are known to perpetuate and amplify harmful soci...
research
08/01/2023

The Bias Amplification Paradox in Text-to-Image Generation

Bias amplification is a phenomenon in which models increase imbalances p...
research
12/02/2019

Exposing and Correcting the Gender Bias in Image Captioning Datasets and Models

The task of image captioning implicitly involves gender identification. ...
research
08/28/2023

Gender bias and stereotypes in Large Language Models

Large Language Models (LLMs) have made substantial progress in the past ...
research
06/14/2018

Neural Stethoscopes: Unifying Analytic, Auxiliary and Adversarial Network Probing

Model interpretability and systematic, targeted model adaptation present...
research
05/12/2022

What's in a Caption? Dataset-Specific Linguistic Diversity and Its Effect on Visual Description Models and Metrics

While there have been significant gains in the field of automated video ...

Please sign up or login with your details

Forgot password? Click here to reset