On Adversarial Bias and the Robustness of Fair Machine Learning

06/15/2020
by   Hongyan Chang, et al.
0

Optimizing prediction accuracy can come at the expense of fairness. Towards minimizing discrimination against a group, fair machine learning algorithms strive to equalize the behavior of a model across different groups, by imposing a fairness constraint on models. However, we show that giving the same importance to groups of different sizes and distributions, to counteract the effect of bias in training data, can be in conflict with robustness. We analyze data poisoning attacks against group-based fair machine learning, with the focus on equalized odds. An adversary who can control sampling or labeling for a fraction of training data, can reduce the test accuracy significantly beyond what he can achieve on unconstrained models. Adversarial sampling and adversarial labeling attacks can also worsen the model's fairness gap on test data, even though the model satisfies the fairness constraint on training data. We analyze the robustness of fair machine learning through an empirical evaluation of attacks on multiple algorithms and benchmark datasets.

READ FULL TEXT

page 34

page 35

research
10/17/2021

Poisoning Attacks on Fair Machine Learning

Both fair machine learning and adversarial learning have been extensivel...
research
11/07/2020

On the Privacy Risks of Algorithmic Fairness

Algorithmic fairness and privacy are essential elements of trustworthy m...
research
05/05/2022

Subverting Fair Image Search with Generative Adversarial Perturbations

In this work we explore the intersection fairness and robustness in the ...
research
02/16/2022

Fairness constraint in Structural Econometrics and Application to fair estimation using Instrumental Variables

A supervised machine learning algorithm determines a model from a learni...
research
11/04/2022

Fairness-aware Regression Robust to Adversarial Attacks

In this paper, we take a first step towards answering the question of ho...
research
07/09/2023

On The Impact of Machine Learning Randomness on Group Fairness

Statistical measures for group fairness in machine learning reflect the ...
research
02/11/2021

Fairness-Aware Learning from Corrupted Data

Addressing fairness concerns about machine learning models is a crucial ...

Please sign up or login with your details

Forgot password? Click here to reset