From Pretraining Data to Language Models to Downstream Tasks: Tracking the Trails of Political Biases Leading to Unfair NLP Models

05/15/2023
by   Shangbin Feng, et al.
0

Large language models (LMs) are pretrained on diverse data sources: news, discussion forums, books, online encyclopedias. A significant portion of this data includes facts and opinions which, on one hand, celebrate democracy and diversity of ideas, and on the other hand are inherently socially biased. Our work develops new methods to (1) measure media biases in LMs trained on such corpora, along the social and economic axes, and (2) measure the fairness of downstream NLP models trained on top of politically biased LMs. We focus on hate speech and misinformation detection, aiming to empirically quantify the effects of political (social, economic) biases in pretraining data on the fairness of high-stakes social-oriented tasks. Our findings reveal that pretrained LMs do have political leanings which reinforce the polarization present in pretraining corpora, propagating social biases into hate speech predictions and media biases into misinformation detectors. We discuss the implications of our findings for NLP research and propose future directions to mitigate the unfairness.

READ FULL TEXT

page 4

page 5

page 6

page 7

research
05/22/2023

Language-Agnostic Bias Detection in Language Models

Pretrained language models (PLMs) are key components in NLP, but they co...
research
12/14/2021

Measuring Fairness with Biased Rulers: A Survey on Quantifying Biases in Pretrained Language Models

An increasing awareness of biased patterns in natural language processin...
research
11/21/2022

Cultural Re-contextualization of Fairness Research in Language Technologies in India

Recent research has revealed undesirable biases in NLP data and models. ...
research
12/29/2022

Political representation bias in DBpedia and Wikidata as a challenge for downstream processing

Diversity Searcher is a tool originally developed to help analyse divers...
research
09/25/2022

Re-contextualizing Fairness in NLP: The Case of India

Recent research has revealed undesirable biases in NLP data and models. ...
research
08/03/2023

A Multidimensional Analysis of Social Biases in Vision Transformers

The embedding spaces of image models have been shown to encode a range o...
research
08/28/2023

AI in the Gray: Exploring Moderation Policies in Dialogic Large Language Models vs. Human Answers in Controversial Topics

The introduction of ChatGPT and the subsequent improvement of Large Lang...

Please sign up or login with your details

Forgot password? Click here to reset