Residual Unfairness in Fair Machine Learning from Prejudiced Data

06/07/2018
by   Nathan Kallus, et al.
0

Recent work in fairness in machine learning has proposed adjusting for fairness by equalizing accuracy metrics across groups and has also studied how datasets affected by historical prejudices may lead to unfair decision policies. We connect these lines of work and study the residual unfairness that arises when a fairness-adjusted predictor is not actually fair on the target population due to systematic censoring of training data by existing biased policies. This scenario is particularly common in the same applications where fairness is a concern. We characterize theoretically the impact of such censoring on standard fairness metrics for binary classifiers and provide criteria for when residual unfairness may or may not appear. We prove that, under certain conditions, fairness-adjusted classifiers will in fact induce residual unfairness that perpetuates the same injustices, against the same groups, that biased the data to begin with, thus showing that even state-of-the-art fair machine learning can have a "bias in, bias out" property. When certain benchmark data is available, we show how sample reweighting can estimate and adjust fairness metrics while accounting for censoring. We use this to study the case of Stop, Question, and Frisk (SQF) and demonstrate that attempting to adjust for fairness perpetuates the same injustices that the policy is infamous for.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/12/2023

On Testing and Comparing Fair classifiers under Data Bias

In this paper, we consider a theoretical model for injecting data bias, ...
research
01/15/2019

Identifying and Correcting Label Bias in Machine Learning

Datasets often contain biases which unfairly disadvantage certain groups...
research
12/21/2022

Consistent Range Approximation for Fair Predictive Modeling

This paper proposes a novel framework for certifying the fairness of pre...
research
10/25/2021

Fair Enough: Searching for Sufficient Measures of Fairness

Testing machine learning software for ethical bias has become a pressing...
research
06/12/2020

Fairness in Forecasting and Learning Linear Dynamical Systems

As machine learning becomes more pervasive, the urgency of assuring its ...
research
01/18/2021

Through the Data Management Lens: Experimental Analysis and Evaluation of Fair Classification

Classification, a heavily-studied data-driven machine learning task, dri...
research
12/04/2022

Fairness in Contextual Resource Allocation Systems: Metrics and Incompatibility Results

We study critical systems that allocate scarce resources to satisfy basi...

Please sign up or login with your details

Forgot password? Click here to reset