Explainability for fair machine learning

10/14/2020
by   Tom Begley, et al.
0

As the decisions made or influenced by machine learning models increasingly impact our lives, it is crucial to detect, understand, and mitigate unfairness. But even simply determining what "unfairness" should mean in a given context is non-trivial: there are many competing definitions, and choosing between them often requires a deep understanding of the underlying task. It is thus tempting to use model explainability to gain insights into model fairness, however existing explainability tools do not reliably indicate whether a model is indeed fair. In this work we present a new approach to explaining fairness in machine learning, based on the Shapley value paradigm. Our fairness explanations attribute a model's overall unfairness to individual input features, even in cases where the model does not operate on sensitive attributes directly. Moreover, motivated by the linearity of Shapley explainability, we propose a meta algorithm for applying existing training-time fairness interventions, wherein one trains a perturbation to the original model, rather than a new model entirely. By explaining the original model, the perturbation, and the fair-corrected model, we gain insight into the accuracy-fairness trade-off that is being made by the intervention. We further show that this meta algorithm enjoys both flexibility and stability benefits with no loss in performance.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/31/2023

Superhuman Fairness

The fairness of machine learning-based decisions has become an increasin...
research
12/07/2022

Fairness and Explainability: Bridging the Gap Towards Fair Model Explanations

While machine learning models have achieved unprecedented success in rea...
research
06/08/2022

Challenges in Applying Explainability Methods to Improve the Fairness of NLP Models

Motivations for methods in explainable artificial intelligence (XAI) oft...
research
11/01/2021

Gradient Frequency Modulation for Visually Explaining Video Understanding Models

In many applications, it is essential to understand why a machine learni...
research
10/24/2019

Fairness Sample Complexity and the Case for Human Intervention

With the aim of building machine learning systems that incorporate stand...
research
11/06/2019

Fair Meta-Learning: Learning How to Learn Fairly

Data sets for fairness relevant tasks can lack examples or be biased acc...
research
10/10/2022

FEAMOE: Fair, Explainable and Adaptive Mixture of Experts

Three key properties that are desired of trustworthy machine learning mo...

Please sign up or login with your details

Forgot password? Click here to reset