Exposing Bias in Online Communities through Large-Scale Language Models

06/04/2023
by   Celine Wald, et al.
0

Progress in natural language generation research has been shaped by the ever-growing size of language models. While large language models pre-trained on web data can generate human-sounding text, they also reproduce social biases and contribute to the propagation of harmful stereotypes. This work utilises the flaw of bias in language models to explore the biases of six different online communities. In order to get an insight into the communities' viewpoints, we fine-tune GPT-Neo 1.3B with six social media datasets. The bias of the resulting models is evaluated by prompting the models with different demographics and comparing the sentiment and toxicity values of these generations. Together, these methods reveal that bias differs in type and intensity for the various models. This work not only affirms how easily bias is absorbed from training data but also presents a scalable method to identify and compare the bias of different datasets or communities. Additionally, the examples generated for this work demonstrate the limitations of using automated sentiment and toxicity classifiers in bias research.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/05/2023

Nationality Bias in Text Generation

Little attention is placed on analyzing nationality bias in language mod...
research
04/15/2022

Identifying and Measuring Token-Level Sentiment Bias in Pre-trained Language Models with Prompts

Due to the superior performance, large-scale pre-trained language models...
research
07/04/2023

Robust Hate Speech Detection in Social Media: A Cross-Dataset Empirical Evaluation

The automatic detection of hate speech online is an active research area...
research
06/08/2023

Bias Against 93 Stigmatized Groups in Masked Language Models and Downstream Sentiment Classification Tasks

The rapid deployment of artificial intelligence (AI) models demands a th...
research
06/23/2022

A Disability Lens towards Biases in GPT-3 Generated Open-Ended Languages

Language models (LM) are becoming prevalent in many language-based appli...
research
07/31/2023

DoDo Learning: DOmain-DemOgraphic Transfer in Language Models for Detecting Abuse Targeted at Public Figures

Public figures receive a disproportionate amount of abuse on social medi...
research
06/07/2023

Soft-prompt Tuning for Large Language Models to Evaluate Bias

Prompting large language models has gained immense popularity in recent ...

Please sign up or login with your details

Forgot password? Click here to reset