Model Debiasing via Gradient-based Explanation on Representation

05/20/2023
by   Jindi Zhang, et al.
0

Machine learning systems produce biased results towards certain demographic groups, known as the fairness problem. Recent approaches to tackle this problem learn a latent code (i.e., representation) through disentangled representation learning and then discard the latent code dimensions correlated with sensitive attributes (e.g., gender). Nevertheless, these approaches may suffer from incomplete disentanglement and overlook proxy attributes (proxies for sensitive attributes) when processing real-world data, especially for unstructured data, causing performance degradation in fairness and loss of useful information for downstream tasks. In this paper, we propose a novel fairness framework that performs debiasing with regard to both sensitive attributes and proxy attributes, which boosts the prediction performance of downstream task models without complete disentanglement. The main idea is to, first, leverage gradient-based explanation to find two model focuses, 1) one focus for predicting sensitive attributes and 2) the other focus for predicting downstream task labels, and second, use them to perturb the latent code that guides the training of downstream task models towards fairness and utility goals. We show empirically that our framework works with both disentangled and non-disentangled representation learning methods and achieves better fairness-accuracy trade-off on unstructured and structured datasets than previous state-of-the-art approaches.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/12/2020

Fairness by Learning Orthogonal Disentangled Representations

Learning discriminative powerful representations is a crucial step for m...
research
05/31/2019

On the Fairness of Disentangled Representations

Recently there has been a significant interest in learning disentangled ...
research
07/09/2021

Multiaccurate Proxies for Downstream Fairness

We study the problem of training a model that must obey demographic fair...
research
12/06/2021

Encouraging Disentangled and Convex Representation with Controllable Interpolation Regularization

We focus on controllable disentangled representation learning (C-Dis-RL)...
research
06/17/2022

Learning Fair Representation via Distributional Contrastive Disentanglement

Learning fair representation is crucial for achieving fairness or debias...
research
08/25/2020

Sensitive Information Detection: Recursive Neural Networks for Encoding Context

The amount of data for processing and categorization grows at an ever in...
research
06/23/2021

Fairness via Representation Neutralization

Existing bias mitigation methods for DNN models primarily work on learni...

Please sign up or login with your details

Forgot password? Click here to reset