Mitigating Shortcuts in Language Models with Soft Label Encoding

09/17/2023
by   Zirui He, et al.
0

Recent research has shown that large language models rely on spurious correlations in the data for natural language understanding (NLU) tasks. In this work, we aim to answer the following research question: Can we reduce spurious correlations by modifying the ground truth labels of the training data? Specifically, we propose a simple yet effective debiasing framework, named Soft Label Encoding (SoftLE). We first train a teacher model with hard labels to determine each sample's degree of relying on shortcuts. We then add one dummy class to encode the shortcut degree, which is used to smooth other dimensions in the ground truth label to generate soft labels. This new ground truth label is used to train a more robust student model. Extensive experiments on two NLU benchmark tasks demonstrate that SoftLE significantly improves out-of-distribution generalization while maintaining satisfactory in-distribution accuracy.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/10/2020

Multitask Emotion Recognition with Incomplete Labels

We train a unified model to perform three tasks: facial action unit dete...
research
02/10/2020

FAU, Facial Expressions, Valence and Arousal: A Multi-task Solution

In the paper, we aim to train a unified model that performs three tasks:...
research
07/13/2022

Beyond Hard Labels: Investigating data label distributions

High-quality data is a key aspect of modern machine learning. However, l...
research
07/01/2021

An Investigation of the (In)effectiveness of Counterfactually Augmented Data

While pretrained language models achieve excellent performance on natura...
research
07/27/2022

NICEST: Noisy Label Correction and Training for Robust Scene Graph Generation

Nearly all existing scene graph generation (SGG) models have overlooked ...
research
09/17/2021

Self-training with Few-shot Rationalization: Teacher Explanations Aid Student in Few-shot NLU

While pre-trained language models have obtained state-of-the-art perform...
research
03/24/2022

Generating Data to Mitigate Spurious Correlations in Natural Language Inference Datasets

Natural language processing models often exploit spurious correlations b...

Please sign up or login with your details

Forgot password? Click here to reset