LUCID-GAN: Conditional Generative Models to Locate Unfairness

07/28/2023
by   Andres Algaba, et al.
0

Most group fairness notions detect unethical biases by computing statistical parity metrics on a model's output. However, this approach suffers from several shortcomings, such as philosophical disagreement, mutual incompatibility, and lack of interpretability. These shortcomings have spurred the research on complementary bias detection methods that offer additional transparency into the sources of discrimination and are agnostic towards an a priori decision on the definition of fairness and choice of protected features. A recent proposal in this direction is LUCID (Locating Unfairness through Canonical Inverse Design), where canonical sets are generated by performing gradient descent on the input space, revealing a model's desired input given a preferred output. This information about the model's mechanisms, i.e., which feature values are essential to obtain specific outputs, allows exposing potential unethical biases in its internal logic. Here, we present LUCID-GAN, which generates canonical inputs via a conditional generative model instead of gradient-based inverse design. LUCID-GAN has several benefits, including that it applies to non-differentiable models, ensures that canonical sets consist of realistic inputs, and allows to assess proxy and intersectional discrimination. We empirically evaluate LUCID-GAN on the UCI Adult and COMPAS data sets and show that it allows for detecting unethical biases in black-box models without requiring access to the training data.

READ FULL TEXT
research
08/26/2022

LUCID: Exposing Algorithmic Bias through Inverse Design

AI systems can create, propagate, support, and automate bias in decision...
research
06/21/2019

FlipTest: Fairness Auditing via Optimal Transport

We present FlipTest, a black-box auditing technique for uncovering subgr...
research
09/02/2022

A Discussion of Discrimination and Fairness in Insurance Pricing

Indirect discrimination is an issue of major concern in algorithmic mode...
research
11/15/2016

Iterative Orthogonal Feature Projection for Diagnosing Bias in Black-Box Models

Predictive models are increasingly deployed for the purpose of determini...
research
03/14/2023

Demographic Parity Inspector: Fairness Audits via the Explanation Space

Even if deployed with the best intentions, machine learning methods can ...
research
06/06/2020

A Generic and Model-Agnostic Exemplar Synthetization Framework for Explainable AI

With the growing complexity of deep learning methods adopted in practica...

Please sign up or login with your details

Forgot password? Click here to reset