Poisoning Attacks on Fair Machine Learning

10/17/2021
by   Minh-Hao Van, et al.
0

Both fair machine learning and adversarial learning have been extensively studied. However, attacking fair machine learning models has received less attention. In this paper, we present a framework that seeks to effectively generate poisoning samples to attack both model accuracy and algorithmic fairness. Our attacking framework can target fair machine learning models trained with a variety of group based fairness notions such as demographic parity and equalized odds. We develop three online attacks, adversarial sampling , adversarial labeling, and adversarial feature modification. All three attacks effectively and efficiently produce poisoning samples via sampling, labeling, or modifying a fraction of training data in order to reduce the test accuracy. Our framework enables attackers to flexibly adjust the attack's focus on prediction accuracy or fairness and accurately quantify the impact of each candidate point to both accuracy loss and fairness violation, thus producing effective poisoning samples. Experiments on two real datasets demonstrate the effectiveness and efficiency of our framework.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/15/2020

On Adversarial Bias and the Robustness of Fair Machine Learning

Optimizing prediction accuracy can come at the expense of fairness. Towa...
research
11/04/2022

Fairness-aware Regression Robust to Adversarial Attacks

In this paper, we take a first step towards answering the question of ho...
research
04/15/2020

Poisoning Attacks on Algorithmic Fairness

Research in adversarial machine learning has shown how the performance o...
research
10/10/2022

FEAMOE: Fair, Explainable and Adaptive Mixture of Experts

Three key properties that are desired of trustworthy machine learning mo...
research
10/18/2022

Towards Fair Classification against Poisoning Attacks

Fair classification aims to stress the classification models to achieve ...
research
10/30/2019

DADI: Dynamic Discovery of Fair Information with Adversarial Reinforcement Learning

We introduce a framework for dynamic adversarial discovery of informatio...
research
05/23/2019

Generative Adversarial Networks for Mitigating Biases in Machine Learning Systems

In this paper, we propose a new framework for mitigating biases in machi...

Please sign up or login with your details

Forgot password? Click here to reset