Mitigating Label Bias via Decoupled Confident Learning

07/18/2023
by   Yunyi Li, et al.
0

Growing concerns regarding algorithmic fairness have led to a surge in methodologies to mitigate algorithmic bias. However, such methodologies largely assume that observed labels in training data are correct. This is problematic because bias in labels is pervasive across important domains, including healthcare, hiring, and content moderation. In particular, human-generated labels are prone to encoding societal biases. While the presence of labeling bias has been discussed conceptually, there is a lack of methodologies to address this problem. We propose a pruning method – Decoupled Confident Learning (DeCoLe) – specifically designed to mitigate label bias. After illustrating its performance on a synthetic dataset, we apply DeCoLe in the context of hate speech detection, where label bias has been recognized as an important challenge, and show that it successfully identifies biased labels and outperforms competing approaches.

READ FULL TEXT

page 4

page 5

page 6

page 7

research
07/15/2022

More Data Can Lead Us Astray: Active Data Acquisition in the Presence of Label Bias

An increased awareness concerning risks of algorithmic bias has driven a...
research
01/10/2022

The Dataset Nutrition Label (2nd Gen): Leveraging Context to Mitigate Harms in Artificial Intelligence

As the production of and reliance on datasets to produce automated decis...
research
02/27/2020

ConQUR: Mitigating Delusional Bias in Deep Q-learning

Delusional bias is a fundamental source of error in approximate Q-learni...
research
06/21/2019

Mitigating Bias in Algorithmic Employment Screening: Evaluating Claims and Practices

There has been rapidly growing interest in the use of algorithms for emp...
research
05/28/2023

Mitigating Label Biases for In-context Learning

Various design settings for in-context learning (ICL), such as the choic...
research
05/23/2023

Mitigating Label Noise through Data Ambiguation

Label noise poses an important challenge in machine learning, especially...
research
05/31/2021

Using Pareto Simulated Annealing to Address Algorithmic Bias in Machine Learning

Algorithmic Bias can be due to bias in the training data or issues with ...

Please sign up or login with your details

Forgot password? Click here to reset