Social Bias Meets Data Bias: The Impacts of Labeling and Measurement Errors on Fairness Criteria

05/31/2022
by   Yiqiao Liao, et al.
0

Although many fairness criteria have been proposed to ensure that machine learning algorithms do not exhibit or amplify our existing social biases, these algorithms are trained on datasets that can themselves be statistically biased. In this paper, we investigate the robustness of a number of existing (demographic) fairness criteria when the algorithm is trained on biased data. We consider two forms of dataset bias: errors by prior decision makers in the labeling process, and errors in measurement of the features of disadvantaged individuals. We analytically show that some constraints (such as Demographic Parity) can remain robust when facing certain statistical biases, while others (such as Equalized Odds) are significantly violated if trained on biased data. We also analyze the sensitivity of these criteria and the decision maker's utility to biases. We provide numerical experiments based on three real-world datasets (the FICO, Adult, and German credit score datasets) supporting our analytical findings. Our findings present an additional guideline for choosing among existing fairness criteria, or for proposing new criteria, when available datasets may be biased.

READ FULL TEXT
research
01/15/2019

Fair and Unbiased Algorithmic Decision Making: Current State and Future Challenges

Machine learning algorithms are now frequently used in sensitive context...
research
06/13/2018

Comparing Fairness Criteria Based on Social Outcome

Fairness in algorithmic decision-making processes is attracting increasi...
research
06/30/2018

Achieving Fairness through Adversarial Learning: an Application to Recidivism Prediction

Recidivism prediction scores are used across the USA to determine senten...
research
12/02/2019

Recovering from Biased Data: Can Fairness Constraints Improve Accuracy?

Multiple fairness constraints have been proposed in the literature, moti...
research
11/17/2017

Predict Responsibly: Increasing Fairness by Learning To Defer

Machine learning systems, which are often used for high-stakes decisions...
research
02/23/2020

Fair Adversarial Networks

The influence of human judgement is ubiquitous in datasets used across t...
research
07/24/2019

On the bias of H-scores for comparing biclusters, and how to correct it

In the last two decades several biclustering methods have been developed...

Please sign up or login with your details

Forgot password? Click here to reset