An Empirical Study of Memorization in NLP

03/23/2022
by   Xiaosen Zheng, et al.
0

A recent study by Feldman (2020) proposed a long-tail theory to explain the memorization behavior of deep learning models. However, memorization has not been empirically verified in the context of NLP, a gap addressed by this work. In this paper, we use three different NLP tasks to check if the long-tail theory holds. Our experiments demonstrate that top-ranked memorized training instances are likely atypical, and removing the top-memorized training instances leads to a more serious drop in test accuracy compared with removing training instances randomly. Furthermore, we develop an attribution method to better understand why a training instance is memorized. We empirically show that our memorization attribution method is faithful, and share our interesting finding that the top-memorized parts of a training instance tend to be features negatively correlated with the class label.

READ FULL TEXT
research
04/09/2021

An Empirical Comparison of Instance Attribution Methods for NLP

Widespread adoption of deep models has motivated a pressing need for app...
research
07/01/2021

Combining Feature and Instance Attribution to Detect Artifacts

Training the large deep neural networks that dominate NLP requires large...
research
11/15/2021

Fast Axiomatic Attribution for Neural Networks

Mitigating the dependence on spurious correlations present in the traini...
research
01/01/2021

On Explaining Your Explanations of BERT: An Empirical Study with Sequence Classification

BERT, as one of the pretrianed language models, attracts the most attent...
research
11/04/2021

Adversarial Attacks on Knowledge Graph Embeddings via Instance Attribution Methods

Despite the widespread use of Knowledge Graph Embeddings (KGE), little i...
research
05/19/2023

Mitigating Backdoor Poisoning Attacks through the Lens of Spurious Correlation

Modern NLP models are often trained over large untrusted datasets, raisi...
research
01/03/2023

Data Valuation Without Training of a Model

Many recent works on understanding deep learning try to quantify how muc...

Please sign up or login with your details

Forgot password? Click here to reset