FAIR: Fair Adversarial Instance Re-weighting

11/15/2020
by   Andrija Petrović, et al.
0

With growing awareness of societal impact of artificial intelligence, fairness has become an important aspect of machine learning algorithms. The issue is that human biases towards certain groups of population, defined by sensitive features like race and gender, are introduced to the training data through data collection and labeling. Two important directions of fairness ensuring research have focused on (i) instance weighting in order to decrease the impact of more biased instances and (ii) adversarial training in order to construct data representations informative of the target variable, but uninformative of the sensitive attributes. In this paper we propose a Fair Adversarial Instance Re-weighting (FAIR) method, which uses adversarial training to learn instance weighting function that ensures fair predictions. Merging the two paradigms, it inherits desirable properties from both – interpretability of reweighting and end-to-end trainability of adversarial training. We propose four different variants of the method and, among other things, demonstrate how the method can be cast in a fully probabilistic framework. Additionally, theoretical analysis of FAIR models' properties have been studied extensively. We compare FAIR models to 7 other related and state-of-the-art models and demonstrate that FAIR is able to achieve a better trade-off between accuracy and unfairness. To the best of our knowledge, this is the first model that merges reweighting and adversarial approaches by means of a weighting function that can provide interpretable information about fairness of individual instances.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/13/2020

To be Robust or to be Fair: Towards Fairness in Adversarial Training

Adversarial training algorithms have been proven to be reliable to impro...
research
03/03/2023

FairShap: A Data Re-weighting Approach for Algorithmic Fairness based on Shapley Values

In this paper, we propose FairShap, a novel and interpretable pre-proces...
research
02/07/2022

Learning fair representation with a parametric integral probability metric

As they have a vital effect on social decision-making, AI algorithms sho...
research
07/18/2022

On Fair Classification with Mostly Private Sensitive Attributes

Machine learning models have demonstrated promising performance in many ...
research
02/02/2022

Normalise for Fairness: A Simple Normalisation Technique for Fairness in Regression Machine Learning Problems

Algorithms and Machine Learning (ML) are increasingly affecting everyday...
research
10/14/2022

InterFair: Debiasing with Natural Language Feedback for Fair Interpretable Predictions

Debiasing methods in NLP models traditionally focus on isolating informa...
research
12/21/2022

Consistent Range Approximation for Fair Predictive Modeling

This paper proposes a novel framework for certifying the fairness of pre...

Please sign up or login with your details

Forgot password? Click here to reset