Exposing Length Divergence Bias of Textual Matching Models

09/06/2021
by   Lan Jiang, et al.
0

Despite the remarkable success deep models have achieved in Textual Matching (TM), their robustness issue is still a topic of concern. In this work, we propose a new perspective to study this issue – via the length divergence bias of TM models. We conclude that this bias stems from two parts: the label bias of existing TM datasets and the sensitivity of TM models to superficial information. We critically examine widely used TM datasets, and find that all of them follow specific length divergence distributions by labels, providing direct cues for predictions. As for the TM models, we conduct adversarial evaluation and show that all models' performances drop on the out-of-distribution adversarial test sets we construct, which demonstrates that they are all misled by biased training sets. This is also confirmed by the SentLen probing task that all models capture rich length information during training to facilitate their performances. Finally, to alleviate the length divergence bias in TM models, we propose a practical adversarial training method using bias-free training data. Our experiments indicate that we successfully improve the robustness and generalization ability of models at the same time.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/15/2019

Selection Bias Explorations and Debias Methods for Natural Language Sentence Matching Datasets

Natural Language Sentence Matching (NLSM) has gained substantial attenti...
research
10/23/2020

Improving Robustness by Augmenting Training Sentences with Predicate-Argument Structures

Existing NLP datasets contain various biases, and models tend to quickly...
research
07/21/2022

AugRmixAT: A Data Processing and Training Method for Improving Multiple Robustness and Generalization Performance

Deep neural networks are powerful, but they also have shortcomings such ...
research
08/26/2022

Lower Difficulty and Better Robustness: A Bregman Divergence Perspective for Adversarial Training

In this paper, we investigate on improving the adversarial robustness ob...
research
07/17/2020

Learning to Match Distributions for Domain Adaptation

When the training and test data are from different distributions, domain...
research
05/29/2021

Exploiting Position Bias for Robust Aspect Sentiment Classification

Aspect sentiment classification (ASC) aims at determining sentiments exp...
research
05/22/2019

Sentence Length

The distribution of sentence length in ordinary language is not well cap...

Please sign up or login with your details

Forgot password? Click here to reset