Investigating Bias in Image Classification using Model Explanations

12/10/2020
by   Schrasing Tong, et al.
0

We evaluated whether model explanations could efficiently detect bias in image classification by highlighting discriminating features, thereby removing the reliance on sensitive attributes for fairness calculations. To this end, we formulated important characteristics for bias detection and observed how explanations change as the degree of bias in models change. The paper identifies strengths and best practices for detecting bias using explanations, as well as three main weaknesses: explanations poorly estimate the degree of bias, could potentially introduce additional bias into the analysis, and are sometimes inefficient in terms of human effort involved.

READ FULL TEXT

page 4

page 5

research
05/05/2020

Global explanations for discovering bias in data

In the paper, we propose attention-based summarized post-hoc explanation...
research
11/06/2020

Wasserstein-based fairness interpretability framework for machine learning models

In this article, we introduce a fairness interpretability framework for ...
research
04/10/2023

Explanation Strategies for Image Classification in Humans vs. Current Explainable AI

Explainable AI (XAI) methods provide explanations of AI models, but our ...
research
07/12/2023

Learning from Exemplary Explanations

eXplanation Based Learning (XBL) is a form of Interactive Machine Learni...
research
12/17/2021

Interpretable Data-Based Explanations for Fairness Debugging

A wide variety of fairness metrics and eXplainable Artificial Intelligen...
research
04/14/2021

Is Disentanglement all you need? Comparing Concept-based Disentanglement Approaches

Concept-based explanations have emerged as a popular way of extracting h...
research
06/11/2020

How Interpretable and Trustworthy are GAMs?

Generalized additive models (GAMs) have become a leading model class for...

Please sign up or login with your details

Forgot password? Click here to reset