Quantifying Feature Contributions to Overall Disparity Using Information Theory

06/16/2022
by   Sanghamitra Dutta, et al.
19

When a machine-learning algorithm makes biased decisions, it can be helpful to understand the sources of disparity to explain why the bias exists. Towards this, we examine the problem of quantifying the contribution of each individual feature to the observed disparity. If we have access to the decision-making model, one potential approach (inspired from intervention-based approaches in explainability literature) is to vary each individual feature (while keeping the others fixed) and use the resulting change in disparity to quantify its contribution. However, we may not have access to the model or be able to test/audit its outputs for individually varying features. Furthermore, the decision may not always be a deterministic function of the input features (e.g., with human-in-the-loop). For these situations, we might need to explain contributions using purely distributional (i.e., observational) techniques, rather than interventional. We ask the question: what is the "potential" contribution of each individual feature to the observed disparity in the decisions when the exact decision-making mechanism is not accessible? We first provide canonical examples (thought experiments) that help illustrate the difference between distributional and interventional approaches to explaining contributions, and when either is better suited. When unable to intervene on the inputs, we quantify the "redundant" statistical dependency about the protected attribute that is present in both the final decision and an individual feature, by leveraging a body of work in information theory called Partial Information Decomposition. We also perform a simple case study to show how this technique could be applied to quantify contributions.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/27/2019

Using Social Choice Theory to Finalize Architectural Decisions

Unbiased and objective architectural design decisions are crucial for th...
research
03/30/2023

Shapley Chains: Extending Shapley Values to Classifier Chains

In spite of increased attention on explainable machine learning models, ...
research
08/19/2022

Personalized Decision Making – A Conceptual Introduction

Personalized decision making targets the behavior of a specific individu...
research
05/25/2023

Bias, Consistency, and Partisanship in U.S. Asylum Cases: A Machine Learning Analysis of Extraneous Factors in Immigration Court Decisions

In this study, we introduce a novel two-pronged scoring system to measur...
research
07/21/2023

Demystifying Local and Global Fairness Trade-offs in Federated Learning Using Partial Information Decomposition

In this paper, we present an information-theoretic perspective to group ...
research
03/03/2023

Model Explanation Disparities as a Fairness Diagnostic

In recent years, there has been a flurry of research focusing on the fai...
research
07/17/2018

Explicating feature contribution using Random Forest proximity distances

In Random Forests, proximity distances are a metric representation of da...

Please sign up or login with your details

Forgot password? Click here to reset