Fairness by Explicability and Adversarial SHAP Learning

03/11/2020
by   James M. Hickey, et al.
0

The ability to understand and trust the fairness of model predictions, particularly when considering the outcomes of unprivileged groups, is critical to the deployment and adoption of machine learning systems. SHAP values provide a unified framework for interpreting model predictions and feature attribution but do not address the problem of fairness directly. In this work, we propose a new definition of fairness that emphasises the role of an external auditor and model explicability. To satisfy this definition, we develop a framework for mitigating model bias using regularizations constructed from the SHAP values of an adversarial surrogate model. We focus on the binary classification task with a single unprivileged group and link our fairness explicability constraints to classical statistical fairness metrics. We demonstrate our approaches using gradient and adaptive boosting on: a synthetic dataset, the UCI Adult (Census) Dataset and a real-world credit scoring dataset. The models produced were fairer and performant.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/25/2020

Counterfactual fairness: removing direct effects through regularization

Building machine learning models that are fair with respect to an unpriv...
research
03/01/2023

Fairness Evaluation in Text Classification: Machine Learning Practitioner Perspectives of Individual and Group Fairness

Mitigating algorithmic bias is a critical task in the development and de...
research
05/14/2020

Statistical Equity: A Fairness Classification Objective

Machine learning systems have been shown to propagate the societal error...
research
10/12/2018

Interpretable Fairness via Target Labels in Gaussian Process Models

Addressing fairness in machine learning models has recently attracted a ...
research
12/04/2022

Fairness in Contextual Resource Allocation Systems: Metrics and Incompatibility Results

We study critical systems that allocate scarce resources to satisfy basi...
research
02/09/2023

On Fairness and Stability: Is Estimator Variance a Friend or a Foe?

The error of an estimator can be decomposed into a (statistical) bias te...
research
10/25/2019

Toward a better trade-off between performance and fairness with kernel-based distribution matching

As recent literature has demonstrated how classifiers often carry uninte...

Please sign up or login with your details

Forgot password? Click here to reset