Memorization vs. Generalization: Quantifying Data Leakage in NLP Performance Evaluation

02/03/2021
by   Aparna Elangovan, et al.
0

Public datasets are often used to evaluate the efficacy and generalizability of state-of-the-art methods for many tasks in natural language processing (NLP). However, the presence of overlap between the train and test datasets can lead to inflated results, inadvertently evaluating the model's ability to memorize and interpreting it as the ability to generalize. In addition, such data sets may not provide an effective indicator of the performance of these methods in real world scenarios. We identify leakage of training data into test data on several publicly available datasets used to evaluate NLP tasks, including named entity recognition and relation extraction, and study them to assess the impact of that leakage on the model's ability to memorize versus generalize.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/14/2023

MatSci-NLP: Evaluating Scientific Language Models on Materials Science Language Tasks Using Text-to-Schema Modeling

We present MatSci-NLP, a natural language benchmark for evaluating the p...
research
05/07/2022

Number Entity Recognition

Numbers are essential components of text, like any other word tokens, fr...
research
07/01/2019

Is artificial data useful for biomedical Natural Language Processing

A major obstacle to the development of Natural Language Processing (NLP)...
research
06/10/2021

CogAlign: Learning to Align Textual Neural Representations to Cognitive Language Processing Signals

Most previous studies integrate cognitive language processing signals (e...
research
03/21/2021

TextFlint: Unified Multilingual Robustness Evaluation Toolkit for Natural Language Processing

Various robustness evaluation methodologies from different perspectives ...
research
06/15/2022

Contextualization and Generalization in Entity and Relation Extraction

During the past decade, neural networks have become prominent in Natural...
research
10/07/2022

Distillation-Resistant Watermarking for Model Protection in NLP

How can we protect the intellectual property of trained NLP models? Mode...

Please sign up or login with your details

Forgot password? Click here to reset