Towards Realistic Single-Task Continuous Learning Research for NER

10/27/2021
by   Justin Payan, et al.
10

There is an increasing interest in continuous learning (CL), as data privacy is becoming a priority for real-world machine learning applications. Meanwhile, there is still a lack of academic NLP benchmarks that are applicable for realistic CL settings, which is a major challenge for the advancement of the field. In this paper we discuss some of the unrealistic data characteristics of public datasets, study the challenges of realistic single-task continuous learning as well as the effectiveness of data rehearsal as a way to mitigate accuracy loss. We construct a CL NER dataset from an existing publicly available dataset and release it along with the code to the research community.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/22/2021

MasakhaNER: Named Entity Recognition for African Languages

We take a step towards addressing the under-representation of the Africa...
research
11/24/2022

Turning the Tables: Biased, Imbalanced, Dynamic Tabular Datasets for ML Evaluation

Evaluating new techniques on realistic datasets plays a crucial role in ...
research
08/06/2021

Lights, Camera, Action! A Framework to Improve NLP Accuracy over OCR documents

Document digitization is essential for the digital transformation of our...
research
01/30/2018

PEYMA: A Tagged Corpus for Persian Named Entities

The goal in the NER task is to classify proper nouns of a text into clas...
research
05/27/2021

Open-world Machine Learning: Applications, Challenges, and Opportunities

Traditional machine learning especially supervised learning follows the ...
research
11/10/2015

USFD: Twitter NER with Drift Compensation and Linked Data

This paper describes a pilot NER system for Twitter, comprising the USFD...
research
09/17/2021

reproducing "ner and pos when nothing is capitalized"

Capitalization is an important feature in many NLP tasks such as Named E...

Please sign up or login with your details

Forgot password? Click here to reset