Towards Equal Gender Representation in the Annotations of Toxic Language Detection

06/04/2021
by   Elizabeth Excell, et al.
10

Classifiers tend to propagate biases present in the data on which they are trained. Hence, it is important to understand how the demographic identities of the annotators of comments affect the fairness of the resulting model. In this paper, we focus on the differences in the ways men and women annotate comments for toxicity, investigating how these differences result in models that amplify the opinions of male annotators. We find that the BERT model as-sociates toxic comments containing offensive words with male annotators, causing the model to predict 67.7 this disparity between gender predictions can be mitigated by removing offensive words and highly toxic comments from the training data. We then apply the learned associations between gender and language to toxic language classifiers, finding that models trained exclusively on female-annotated data perform 1.8 training models on data after removing all offensive words reduces bias in the model by 55.5

READ FULL TEXT

page 5

page 7

research
06/29/2020

Reading Between the Demographic Lines: Resolving Sources of Bias in Toxicity Classifiers

The censorship of toxic comments is often left to the judgment of imperf...
research
07/05/2017

Like trainer, like bot? Inheritance of bias in algorithmic content moderation

The internet has become a central medium through which `networked public...
research
05/01/2022

Is Your Toxicity My Toxicity? Exploring the Impact of Rater Identity on Toxicity Annotation

Machine learning models are commonly used to detect toxicity in online c...
research
09/07/2021

Hi, my name is Martha: Using names to measure and mitigate bias in generative dialogue models

All AI models are susceptible to learning biases in data that they are t...
research
08/05/2023

Elucidate Gender Fairness in Singing Voice Transcription

It is widely known that males and females typically possess different so...
research
01/29/2021

Challenges in Automated Debiasing for Toxic Language Detection

Biased associations have been a challenge in the development of classifi...
research
08/30/2019

Automatically Inferring Gender Associations from Language

In this paper, we pose the question: do people talk about women and men ...

Please sign up or login with your details

Forgot password? Click here to reset