Biased Models Have Biased Explanations

12/20/2020
by   Aditya Jain, et al.
0

We study fairness in Machine Learning (FairML) through the lens of attribute-based explanations generated for machine learning models. Our hypothesis is: Biased Models have Biased Explanations. To establish that, we first translate existing statistical notions of group fairness and define these notions in terms of explanations given by the model. Then, we propose a novel way of detecting (un)fairness for any black box model. We further look at post-processing techniques for fairness and reason how explanations can be used to make a bias mitigation technique more individually fair. We also introduce a novel post-processing mitigation technique which increases individual fairness in recourse while maintaining group level fairness.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/14/2018

Bias Mitigation Post-processing for Individual and Group Fairness

Whereas previous post-processing approaches for increasing the fairness ...
research
01/31/2022

Fair Wrapping for Black-box Predictions

We introduce a new family of techniques to post-process ("wrap") a black...
research
11/19/2021

Model-agnostic bias mitigation methods with regressor distribution control for Wasserstein-based fairness metrics

This article is a companion paper to our earlier work Miroshnikov et al....
research
05/30/2022

Fooling SHAP with Stealthily Biased Sampling

SHAP explanations aim at identifying which features contribute the most ...
research
06/26/2023

Fairness Aware Counterfactuals for Subgroups

In this work, we present Fairness Aware Counterfactuals for Subgroups (F...
research
02/22/2023

Uncovering Bias in Face Generation Models

Recent advancements in GANs and diffusion models have enabled the creati...
research
06/21/2020

Verifying Individual Fairness in Machine Learning Models

We consider the problem of whether a given decision model, working with ...

Please sign up or login with your details

Forgot password? Click here to reset