Benign Shortcut for Debiasing: Fair Visual Recognition via Intervention with Shortcut Features

08/13/2023
by   Yi Zhang, et al.
0

Machine learning models often learn to make predictions that rely on sensitive social attributes like gender and race, which poses significant fairness risks, especially in societal applications, such as hiring, banking, and criminal justice. Existing work tackles this issue by minimizing the employed information about social attributes in models for debiasing. However, the high correlation between target task and these social attributes makes learning on the target task incompatible with debiasing. Given that model bias arises due to the learning of bias features (i.e., gender) that help target task optimization, we explore the following research question: Can we leverage shortcut features to replace the role of bias feature in target task optimization for debiasing? To this end, we propose Shortcut Debiasing, to first transfer the target task's learning of bias attributes from bias features to shortcut features, and then employ causal intervention to eliminate shortcut features during inference. The key idea of Shortcut Debiasing is to design controllable shortcut features to on one hand replace bias features in contributing to the target task during the training stage, and on the other hand be easily removed by intervention during the inference stage. This guarantees the learning of the target task does not hinder the elimination of bias features. We apply Shortcut Debiasing to several benchmark datasets, and achieve significant improvements over the state-of-the-art debiasing methods in both accuracy and fairness.

READ FULL TEXT

page 3

page 4

research
11/02/2022

Fair Visual Recognition via Intervention with Proxy Features

Deep learning models often learn to make predictions that rely on sensit...
research
06/23/2021

Fairness via Representation Neutralization

Existing bias mitigation methods for DNN models primarily work on learni...
research
08/25/2021

Social Norm Bias: Residual Harms of Fairness-Aware Algorithms

Many modern learning algorithms mitigate bias by enforcing fairness acro...
research
06/08/2018

Blind Justice: Fairness with Encrypted Sensitive Attributes

Recent work has explored how to train machine learning models which do n...
research
08/30/2020

Adversarial Learning for Counterfactual Fairness

In recent years, fairness has become an important topic in the machine l...
research
02/17/2022

Gradient Based Activations for Accurate Bias-Free Learning

Bias mitigation in machine learning models is imperative, yet challengin...
research
10/26/2022

FairCLIP: Social Bias Elimination based on Attribute Prototype Learning and Representation Neutralization

The Vision-Language Pre-training (VLP) models like CLIP have gained popu...

Please sign up or login with your details

Forgot password? Click here to reset