Gender bias and stereotypes in Large Language Models

08/28/2023
by   Hadas Kotek, et al.
0

Large Language Models (LLMs) have made substantial progress in the past several months, shattering state-of-the-art benchmarks in many domains. This paper investigates LLMs' behavior with respect to gender stereotypes, a known issue for prior models. We use a simple paradigm to test the presence of gender bias, building on but differing from WinoBias, a commonly used gender bias dataset, which is likely to be included in the training data of current LLMs. We test four recently published LLMs and demonstrate that they express biased assumptions about men and women's occupations. Our contributions in this paper are as follows: (a) LLMs are 3-6 times more likely to choose an occupation that stereotypically aligns with a person's gender; (b) these choices align with people's perceptions better than with the ground truth as reflected in official job statistics; (c) LLMs in fact amplify the bias beyond what is reflected in perceptions or the ground truth; (d) LLMs ignore crucial ambiguities in sentence structure 95 prompted, they recognize the ambiguity; (e) LLMs provide explanations for their choices that are factually inaccurate and likely obscure the true reason behind their predictions. That is, they provide rationalizations of their biased behavior. This highlights a key property of these models: LLMs are trained on imbalanced datasets; as such, even with the recent successes of reinforcement learning with human feedback, they tend to reflect those imbalances back at us. As with other types of societal biases, we suggest that LLMs must be carefully tested to ensure that they treat minoritized individuals and communities equitably.

READ FULL TEXT
research
05/19/2022

Towards Understanding Gender-Seniority Compound Bias in Natural Language Generation

Women are often perceived as junior to their male counterparts, even wit...
research
05/12/2021

Evaluating Gender Bias in Natural Language Inference

Gender-bias stereotypes have recently raised significant ethical concern...
research
09/18/2023

Bias of AI-Generated Content: An Examination of News Produced by Large Language Models

Large language models (LLMs) have the potential to transform our lives a...
research
03/26/2018

Women also Snowboard: Overcoming Bias in Captioning Models

Most machine learning methods are known to capture and exploit biases of...
research
07/02/2018

Women also Snowboard: Overcoming Bias in Captioning Models (Extended Abstract)

Most machine learning methods are known to capture and exploit biases of...
research
08/23/2021

Exploring Biases and Prejudice of Facial Synthesis via Semantic Latent Space

Deep learning (DL) models are widely used to provide a more convenient a...
research
12/04/2020

Biased Programmers? Or Biased Data? A Field Experiment in Operationalizing AI Ethics

Why do biased predictions arise? What interventions can prevent them? We...

Please sign up or login with your details

Forgot password? Click here to reset