Intersectional Bias in Hate Speech and Abusive Language Datasets

05/12/2020
by   Jae Yeon Kim, et al.
0

Algorithms are widely applied to detect hate speech and abusive language in social media. We investigated whether the human-annotated data used to train these algorithms are biased. We utilized a publicly available annotated Twitter dataset (Founta et al. 2018) and classified the racial, gender, and party identification dimensions of 99,996 tweets. The results showed that African American tweets were up to 3.7 times more likely to be labeled as abusive, and African American male tweets were up to 77 hateful compared to the others. These patterns were statistically significant and robust even when party identification was added as a control variable. This study provides the first systematic evidence on intersectional bias in datasets of hate speech and abusive language.

READ FULL TEXT
research
05/29/2019

Racial Bias in Hate Speech and Abusive Language Detection Datasets

Technologies for abusive language detection are being developed and appl...
research
03/11/2017

Automated Hate Speech Detection and the Problem of Offensive Language

A key challenge for automatic hate-speech detection on social media is t...
research
08/01/2022

Parsimonious Argument Annotations for Hate Speech Counter-narratives

We present an enrichment of the Hateval corpus of hate speech tweets (Ba...
research
08/14/2020

Hate Speech Detection and Racial Bias Mitigation in Social Media based on BERT model

Disparate biases associated with datasets and trained classifiers in hat...
research
11/16/2020

Examining the Feasibility of Off-the-Shelf Algorithms for Masking Directly Identifiable Information in Social Media Data

The identification and removal/replacement of protected information from...
research
07/12/2021

Hate versus Politics: Detection of Hate against Policy makers in Italian tweets

Accurate detection of hate speech against politicians, policy making and...

Please sign up or login with your details

Forgot password? Click here to reset