On Testing and Comparing Fair classifiers under Data Bias

02/12/2023
by   Mohit Sharma, et al.
2

In this paper, we consider a theoretical model for injecting data bias, namely, under-representation and label bias (Blum Stangl, 2019). We theoretically and empirically study its effect on the accuracy and fairness of fair classifiers. Theoretically, we prove that the Bayes optimal group-aware fair classifier on the original data distribution can be recovered by simply minimizing a carefully chosen reweighed loss on the bias-injected distribution. Through extensive experiments on both synthetic and real-world datasets (e.g., Adult, German Credit, Bank Marketing, COMPAS), we empirically audit pre-, in-, and post-processing fair classifiers from standard fairness toolkits for their fairness and accuracy by injecting varying amounts of under-representation and label bias in their training data (but not the test data). Our main observations are: (1) The fairness and accuracy of many standard fair classifiers degrade severely as the bias injected in their training data increases, (2) A simple logistic regression model trained on the right data can often outperform, in both accuracy and fairness, most fair classifiers trained on biased training data, and (3) A few, simple fairness techniques (e.g., reweighing, exponentiated gradients) seem to offer stable accuracy and fairness guarantees even when their training data is injected with under-representation and label bias. Our experiments also show how to integrate a measure of data bias risk in the existing fairness dashboards for real-world deployments

READ FULL TEXT

page 7

page 8

page 21

page 22

page 23

research
12/14/2018

Bias Mitigation Post-processing for Individual and Group Fairness

Whereas previous post-processing approaches for increasing the fairness ...
research
07/12/2020

Ensuring Fairness Beyond the Training Data

We initiate the study of fair classifiers that are robust to perturbatio...
research
07/07/2021

Bias-Tolerant Fair Classification

The label bias and selection bias are acknowledged as two reasons in dat...
research
09/18/2020

Group Fairness by Probabilistic Modeling with Latent Fair Decisions

Machine learning systems are increasingly being used to make impactful d...
research
06/07/2018

Residual Unfairness in Fair Machine Learning from Prejudiced Data

Recent work in fairness in machine learning has proposed adjusting for f...
research
11/29/2021

Learning Fair Classifiers with Partially Annotated Group Labels

Recently, fairness-aware learning have become increasingly crucial, but ...
research
03/10/2019

Fair Logistic Regression: An Adversarial Perspective

Fair prediction methods have primarily been built around existing classi...

Please sign up or login with your details

Forgot password? Click here to reset