The Woman Worked as a Babysitter: On Biases in Language Generation

09/03/2019
by   Emily Sheng, et al.
0

We present a systematic study of biases in natural language generation (NLG) by analyzing text generated from prompts that contain mentions of different demographic groups. In this work, we introduce the notion of the regard towards a demographic, use the varying levels of regard towards different demographics as a defining metric for bias in NLG, and analyze the extent to which sentiment scores are a relevant proxy metric for regard. To this end, we collect strategically-generated text from language models and manually annotate the text with both sentiment and regard scores. Additionally, we build an automatic regard classifier through transfer learning, so that we can analyze biases in unseen text. Together, these methods reveal the extent of the biased nature of language model generations. Our analysis provides a study of biases in NLG, bias metrics and correlated human judgments, and empirical evidence on the usefulness of our annotated dataset.

READ FULL TEXT
research
05/01/2020

Towards Controllable Biases in Language Generation

We present a general approach towards controllable societal biases in na...
research
06/30/2023

Queer People are People First: Deconstructing Sexual Identity Stereotypes in Large Language Models

Large Language Models (LLMs) are trained primarily on minimally processe...
research
04/06/2023

Uncurated Image-Text Datasets: Shedding Light on Demographic Bias

The increasing tendency to collect large and uncurated datasets to train...
research
02/28/2022

Rethinking and Refining the Distinct Metric

Distinct is a widely used automatic metric for evaluating the diversity ...
research
04/18/2021

Revealing Persona Biases in Dialogue Systems

Dialogue systems in the form of chatbots and personal assistants are bei...
research
02/14/2023

A Friendly Face: Do Text-to-Image Systems Rely on Stereotypes when the Input is Under-Specified?

As text-to-image systems continue to grow in popularity with the general...
research
02/03/2021

BiasFinder: Metamorphic Test Generation to Uncover Bias for Sentiment Analysis Systems

Artificial Intelligence (AI) software systems, such as Sentiment Analysi...

Please sign up or login with your details

Forgot password? Click here to reset