Less is Better: Recovering Intended-Feature Subspace to Robustify NLU Models

09/16/2022
by   Ting Wu, et al.
0

Datasets with significant proportions of bias present threats for training a trustworthy model on NLU tasks. Despite yielding great progress, current debiasing methods impose excessive reliance on the knowledge of bias attributes. Definition of the attributes, however, is elusive and varies across different datasets. Furthermore, leveraging these attributes at input level to bias mitigation may leave a gap between intrinsic properties and the underlying decision rule. To narrow down this gap and liberate the supervision on bias, we suggest extending bias mitigation into feature space. Therefore, a novel model, Recovering Intended-Feature Subspace with Knowledge-Free (RISK) is developed. Assuming that shortcut features caused by various biases are unintended for prediction, RISK views them as redundant features. When delving into a lower manifold to remove redundancies, RISK reveals that an extremely low-dimensional subspace with intended features can robustly represent the highly biased dataset. Empirical results demonstrate our model can consistently improve model generalization to out-of-distribution set, and achieves a new state-of-the-art performance.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/30/2022

Bias Mimicking: A Simple Sampling Approach for Bias Mitigation

Prior work has shown that Visual Recognition datasets frequently under-r...
research
07/03/2021

Learning Debiased Representation via Disentangled Feature Augmentation

Image classification models tend to make decisions based on peripheral a...
research
06/30/2021

Fair Visual Recognition in Limited Data Regime using Self-Supervision and Self-Distillation

Deep learning models generally learn the biases present in the training ...
research
02/13/2023

Parameter-efficient Modularised Bias Mitigation via AdapterFusion

Large pre-trained language models contain societal biases and carry alon...
research
11/20/2022

Conceptor-Aided Debiasing of Contextualized Embeddings

Pre-trained language models reflect the inherent social biases of their ...
research
02/17/2022

Gradient Based Activations for Accurate Bias-Free Learning

Bias mitigation in machine learning models is imperative, yet challengin...

Please sign up or login with your details

Forgot password? Click here to reset