Towards Fair Classification against Poisoning Attacks

10/18/2022
by   Han Xu, et al.
5

Fair classification aims to stress the classification models to achieve the equality (treatment or prediction quality) among different sensitive groups. However, fair classification can be under the risk of poisoning attacks that deliberately insert malicious training samples to manipulate the trained classifiers' performance. In this work, we study the poisoning scenario where the attacker can insert a small fraction of samples into training data, with arbitrary sensitive attributes as well as other predictive features. We demonstrate that the fairly trained classifiers can be greatly vulnerable to such poisoning attacks, with much worse accuracy fairness trade-off, even when we apply some of the most effective defenses (originally proposed to defend traditional classification tasks). As countermeasures to defend fair classification tasks, we propose a general and theoretically guaranteed framework which accommodates traditional defense methods to fair classification against poisoning attacks. Through extensive experiments, the results validate that the proposed defense framework obtains better robustness in terms of accuracy and fairness than representative baseline methods.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/30/2022

Learning Fair Models without Sensitive Attributes: A Generative Approach

Most existing fair classifiers rely on sensitive attributes to achieve f...
research
07/12/2020

Ensuring Fairness Beyond the Training Data

We initiate the study of fair classifiers that are robust to perturbatio...
research
06/22/2021

FLEA: Provably Fair Multisource Learning from Unreliable Training Data

Fairness-aware learning aims at constructing classifiers that not only m...
research
10/17/2021

Poisoning Attacks on Fair Machine Learning

Both fair machine learning and adversarial learning have been extensivel...
research
06/10/2021

Fair Classification with Adversarial Perturbations

We study fair classification in the presence of an omniscient adversary ...
research
08/21/2020

Defending Regression Learners Against Poisoning Attacks

Regression models, which are widely used from engineering applications t...
research
04/12/2022

Breaking Fair Binary Classification with Optimal Flipping Attacks

Minimizing risk with fairness constraints is one of the popular approach...

Please sign up or login with your details

Forgot password? Click here to reset