Gender Bias in BERT – Measuring and Analysing Biases through Sentiment Rating in a Realistic Downstream Classification Task

06/27/2023
by   Sophie Jentzsch, et al.
0

Pretrained language models are publicly available and constantly finetuned for various real-life applications. As they become capable of grasping complex contextual information, harmful biases are likely increasingly intertwined with those models. This paper analyses gender bias in BERT models with two main contributions: First, a novel bias measure is introduced, defining biases as the difference in sentiment valuation of female and male sample versions. Second, we comprehensively analyse BERT's biases on the example of a realistic IMDB movie classifier. By systematically varying elements of the training pipeline, we can conclude regarding their impact on the final model bias. Seven different public BERT models in nine training conditions, i.e. 63 models in total, are compared. Almost all conditions yield significant gender biases. Results indicate that reflected biases stem from public BERT models rather than task-specific data, emphasising the weight of responsible usage.

READ FULL TEXT
research
06/18/2019

Measuring Bias in Contextualized Word Representations

Contextual word embeddings such as BERT have achieved state of the art p...
research
06/23/2022

Towards WinoQueer: Developing a Benchmark for Anti-Queer Bias in Large Language Models

This paper presents exploratory work on whether and to what extent biase...
research
11/25/2022

An Analysis of Social Biases Present in BERT Variants Across Multiple Languages

Although large pre-trained language models have achieved great success i...
research
09/10/2020

Investigating Gender Bias in BERT

Contextual language models (CLMs) have pushed the NLP benchmarks to a ne...
research
09/08/2021

Sustainable Modular Debiasing of Language Models

Unfair stereotypical biases (e.g., gender, racial, or religious biases) ...
research
05/31/2022

Hollywood Identity Bias Dataset: A Context Oriented Bias Analysis of Movie Dialogues

Movies reflect society and also hold power to transform opinions. Social...
research
11/15/2021

Assessing gender bias in medical and scientific masked language models with StereoSet

NLP systems use language models such as Masked Language Models (MLMs) th...

Please sign up or login with your details

Forgot password? Click here to reset