Counterfactual Situation Testing: Uncovering Discrimination under Fairness given the Difference

02/23/2023
by   Jose M. Alvarez, et al.
0

We present counterfactual situation testing (CST), a causal data mining framework for detecting discrimination in classifiers. CST aims to answer in an actionable and meaningful way the intuitive question "what would have been the model outcome had the individual, or complainant, been of a different protected status?" It extends the legally-grounded situation testing of Thanh et al. (2011) by operationalizing the notion of fairness given the difference using counterfactual reasoning. For any complainant, we find and compare similar protected and non-protected instances in the dataset used by the classifier to construct a control and test group, where a difference between the decision outcomes of the two groups implies potential individual discrimination. Unlike situation testing, which builds both groups around the complainant, we build the test group on the complainant's counterfactual generated using causal knowledge. The counterfactual is intended to reflect how the protected attribute when changed affects the seemingly neutral attributes used by the classifier, which is taken for granted in many frameworks for discrimination. Under CST, we compare similar individuals within each group but dissimilar individuals across both groups due to the possible difference between the complainant and its counterfactual. Evaluating our framework on two classification scenarios, we show that it uncovers a greater number of cases than situation testing, even when the classifier satisfies the counterfactual fairness condition of Kusner et al. (2017).

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/09/2022

Prediction Sensitivity: Continual Audit of Counterfactual Fairness in Deployed Classifiers

As AI-based systems increasingly impact many areas of our lives, auditin...
research
11/27/2022

"Explain it in the Same Way!" – Model-Agnostic Group Fairness of Counterfactual Explanations

Counterfactual explanations are a popular type of explanation for making...
research
06/21/2019

FlipTest: Fairness Auditing via Optimal Transport

We present FlipTest, a black-box auditing technique for uncovering subgr...
research
06/02/2020

What's Sex Got To Do With Machine Learning

Debate about fairness in machine learning has largely centered around co...
research
06/02/2020

What's Sex Got To Do With Fair Machine Learning?

Debate about fairness in machine learning has largely centered around co...
research
04/09/2023

Information-Theoretic Testing and Debugging of Fairness Defects in Deep Neural Networks

The deep feedforward neural networks (DNNs) are increasingly deployed in...
research
07/10/2020

A Causal Linear Model to Quantify Edge Unfairness for Unfair Edge Prioritization and Discrimination Removal

The dataset can be generated by an unfair mechanism in numerous settings...

Please sign up or login with your details

Forgot password? Click here to reset