On Adversarial Removal of Hypothesis-only Bias in Natural Language Inference

07/09/2019
by   Yonatan Belinkov, et al.
0

Popular Natural Language Inference (NLI) datasets have been shown to be tainted by hypothesis-only biases. Adversarial learning may help models ignore sensitive biases and spurious correlations in data. We evaluate whether adversarial learning can be used in NLI to encourage models to learn representations free of hypothesis-only biases. Our analyses indicate that the representations learned via adversarial learning may be less biased, with only small drops in NLI accuracy.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/09/2019

Don't Take the Premise for Granted: Mitigating Artifacts in Natural Language Inference

Natural Language Inference (NLI) datasets often contain hypothesis-only ...
research
01/25/2021

Diverse Adversaries for Mitigating Bias in Training

Adversarial learning can learn fairer and less biased models of language...
research
04/16/2020

There is Strength in Numbers: Avoiding the Hypothesis-Only Bias in Natural Language Inference via Ensemble Adversarial Training

Natural Language Inference (NLI) datasets contain annotation artefacts r...
research
12/02/2020

Learning from others' mistakes: Avoiding dataset biases without modeling them

State-of-the-art natural language processing (NLP) models often learn to...
research
02/07/2016

The IMP game: Learnability, approximability and adversarial learning beyond Σ^0_1

We introduce a problem set-up we call the Iterated Matching Pennies (IMP...
research
02/10/2020

Adversarial Filters of Dataset Biases

Large neural models have demonstrated human-level performance on languag...
research
12/16/2021

Automatically Identifying Semantic Bias in Crowdsourced Natural Language Inference Datasets

Natural language inference (NLI) is an important task for producing usef...

Please sign up or login with your details

Forgot password? Click here to reset