Neural Styling for Interpretable Fair Representations

10/15/2018
by   Novi Quadrianto, et al.
0

We observe a rapid increase in machine learning models for learning data representations that remove the semantics of protected characteristics, and are therefore able to mitigate unfair prediction outcomes. This is indeed a positive proliferation. All available models however learn latent embeddings, therefore the produced representations do not have the semantic meaning of the input. Our aim here is to learn fair representations that are directly interpretable in the original input domain. We cast this problem as a data-to-data translation; to learn a mapping from data in a source domain to a target domain such that data in the target domain enforces fairness definitions, such as statistical parity or equality of opportunity. Unavailability of fair data in the target domain is the crux of the problem. This paper provides the first approach to learn a highly unconstrained mapping from source to target by maximizing (conditional) dependence of residuals - the difference between data and its translated version - and protected characteristics. The usage of residual statistics ensures that our generated fair data should only be an adjustment of the input data, and this adjustment should reveal the main difference between protected characteristic groups. When applied to CelebA face image dataset with gender as protected characteristic, our model enforces equality of opportunity by adjusting eyes and lips regions. In Adult income dataset, also with gender as protected characteristic, our model achieves equality of opportunity by, among others, obfuscating wife and husband relationship. Visualizing those systematic changes will allow us to scrutinize the interplay of fairness criterion, chosen protected characteristics, and the prediction performance.

READ FULL TEXT
research
04/30/2019

Learning Fair Representations via an Adversarial Framework

Fairness has become a central issue for our research community as classi...
research
07/31/2018

The Measure and Mismeasure of Fairness: A Critical Review of Fair Machine Learning

In one broad class of supervised machine learning problems, researchers ...
research
08/12/2020

Null-sampling for Interpretable and Fair Representations

We propose to learn invariant representations, in the data domain, to ac...
research
06/27/2019

Learning Fair Representations for Kernel Models

Fair representations are a powerful tool for establishing criteria like ...
research
07/19/2019

Fair quantile regression

Quantile regression is a tool for learning conditional distributions. In...
research
10/28/2019

Learning Fair and Interpretable Representations via Linear Orthogonalization

To reduce human error and prejudice, many high-stakes decisions have bee...
research
05/29/2023

Generalized Disparate Impact for Configurable Fairness Solutions in ML

We make two contributions in the field of AI fairness over continuous pr...

Please sign up or login with your details

Forgot password? Click here to reset