Improving the Robustness of Summarization Models by Detecting and Removing Input Noise

12/20/2022
by   Kundan Krishna, et al.
0

The evaluation of abstractive summarization models typically uses test data that is identically distributed as training data. In real-world practice, documents to be summarized may contain input noise caused by text extraction artifacts or data pipeline bugs. The robustness of model performance under distribution shift caused by such noise is relatively under-studied. We present a large empirical study quantifying the sometimes severe loss in performance (up to 12 ROUGE-1 points) from different types of input noise for a range of datasets and model sizes. We then propose a light-weight method for detecting and removing such noise in the input during model inference without requiring any extra training, auxiliary models, or even prior knowledge of the type of noise. Our proposed approach effectively mitigates the loss in performance, recovering a large fraction of the performance drop, sometimes as large as 11 ROUGE-1 points.

READ FULL TEXT

page 2

page 7

research
08/01/2023

Tackling Hallucinations in Neural Chart Summarization

Hallucinations in text generation occur when the system produces text th...
research
06/01/2023

Improving the Robustness of Summarization Systems with Dual Augmentation

A robust summarization system should be able to capture the gist of the ...
research
04/18/2021

Contrastive Out-of-Distribution Detection for Pretrained Transformers

Pretrained transformers achieve remarkable performance when the test dat...
research
10/11/2022

Tackling Instance-Dependent Label Noise with Dynamic Distribution Calibration

Instance-dependent label noise is realistic but rather challenging, wher...
research
07/24/2023

Investigating the Robustness of Sequential Recommender Systems Against Training Data Perturbations: an Empirical Study

Sequential Recommender Systems (SRSs) have been widely used to model use...
research
05/30/2019

On the Accuracy of Influence Functions for Measuring Group Effects

Influence functions estimate the effect of removing particular training ...
research
02/07/2019

Unsupervised Data Uncertainty Learning in Visual Retrieval Systems

We introduce an unsupervised formulation to estimate heteroscedastic unc...

Please sign up or login with your details

Forgot password? Click here to reset