Reducing Gender Bias in Abusive Language Detection

08/22/2018
by   Ji Ho Park, et al.
0

Abusive language detection models tend to have a problem of being biased toward identity words of a certain group of people because of imbalanced training datasets. For example, "You are a good woman" was considered "sexist" when trained on an existing dataset. Such model bias is an obstacle for models to be robust enough for practical use. In this work, we measure gender biases on models trained with different abusive language datasets, while analyzing the effect of different pre-trained word embeddings and model architectures. We also experiment with three bias mitigation methods: (1) debiased word embeddings, (2) gender swap data augmentation, and (3) fine-tuning with a larger corpus. These methods can effectively reduce gender bias by 90-98 can be extended to correct model bias in other scenarios.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/03/2019

Gender-preserving Debiasing for Pre-trained Word Embeddings

Word embeddings learnt from massive text collections have demonstrated s...
research
04/05/2019

Identifying and Reducing Gender Bias in Word-Level Language Models

Many text corpora exhibit socially problematic biases, which can be prop...
research
09/02/2019

It's All in the Name: Mitigating Gender Bias with Name-Based Counterfactual Data Substitution

This paper treats gender bias latent in word embeddings. Previous mitiga...
research
11/14/2022

Does Debiasing Inevitably Degrade the Model Performance

Gender bias in language models has attracted sufficient attention becaus...
research
10/26/2022

A Robust Bias Mitigation Procedure Based on the Stereotype Content Model

The Stereotype Content model (SCM) states that we tend to perceive minor...
research
07/21/2019

Using Word Embeddings to Examine Gender Bias in Dutch Newspapers, 1950-1990

Contemporary debates on filter bubbles and polarization in public and so...
research
07/21/2022

The Birth of Bias: A case study on the evolution of gender bias in an English language model

Detecting and mitigating harmful biases in modern language models are wi...

Please sign up or login with your details

Forgot password? Click here to reset