How True is GPT-2? An Empirical Analysis of Intersectional Occupational Biases

02/08/2021
by   Hannah Kirk, et al.
1

The capabilities of natural language models trained on large-scale data have increased immensely over the past few years. Downstream applications are at risk of inheriting biases contained in these models, with potential negative consequences especially for marginalized groups. In this paper, we analyze the occupational biases of a popular generative language model, GPT-2, intersecting gender with five protected categories: religion, sexuality, ethnicity, political affiliation, and name origin. Using a novel data collection pipeline we collect 396k sentence completions of GPT-2 and find: (i) The machine-predicted jobs are less diverse and more stereotypical for women than for men, especially for intersections; (ii) Fitting 262 logistic models shows intersectional interactions to be highly relevant for occupational associations; (iii) For a given job, GPT-2 reflects the societal skew of gender and ethnicity in the US, and in some cases, pulls the distribution towards gender parity, raising the normative question of what language models _should_ learn.

READ FULL TEXT

page 15

page 18

page 19

page 20

page 24

page 25

page 26

page 29

research
07/28/2020

Defining and Evaluating Fair Natural Language Generation

Our work focuses on the biases that emerge in the natural language gener...
research
08/03/2022

Large scale analysis of gender bias and sexism in song lyrics

We employ Natural Language Processing techniques to analyse 377808 Engli...
research
04/22/2023

Transcending the "Male Code": Implicit Masculine Biases in NLP Contexts

Critical scholarship has elevated the problem of gender bias in data set...
research
09/24/2022

Moral Mimicry: Large Language Models Produce Moral Rationalizations Tailored to Political Identity

Large Language Models (LLMs) have recently demonstrated impressive capab...
research
08/24/2019

Release Strategies and the Social Impacts of Language Models

Large language models have a range of beneficial uses: they can assist i...
research
07/17/2023

Gender mobility in the labor market with skills-based matching models

Skills-based matching promises mobility of workers between different sec...
research
09/11/2023

Detecting Natural Language Biases with Prompt-based Learning

In this project, we want to explore the newly emerging field of prompt e...

Please sign up or login with your details

Forgot password? Click here to reset