A Rigourous Study on Named Entity Recognition: Can Fine-tuning Pretrained Model Lead to the Promised Land?

04/25/2020
by   Hongyu Lin, et al.
0

Fine-tuning pretrained model has achieved promising performance on standard NER benchmarks. Generally, these benchmarks are blessed with strong name regularity, high mention coverage and sufficient context diversity. Unfortunately, when scaling NER to open situations, these advantages may no longer exist, and therefore raise the critical question of whether pretrained supervised models can still work well when facing these issues. As there is no currently available dataset to investigate this problem, this paper proposes to conduct randomization test on standard benchmarks. Specifically, we erase name regularity, mention coverage and context diversity respectively from the benchmarks, in order to explore their impact on the generalization ability of models. Moreover, we also construct a new open NER dataset that focuses on entity types with weak name regularity such as book, song, and movie. From both randomization test and empirical experiments, we draw the conclusions that 1) name regularity is vital for generalization to unseen mentions; 2) high mention coverage may undermine the model generalization ability and 3) context patterns may not require enormous data to capture when using pretrained supervised models.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/09/2022

Domain Generalization using Pretrained Models without Fine-tuning

Fine-tuning pretrained models is a common practice in domain generalizat...
research
05/18/2023

Learning In-context Learning for Named Entity Recognition

Named entity recognition in real-world applications suffers from the div...
research
02/21/2019

Pretrained language model transfer on neural named entity recognition in Indonesian conversational texts

Named entity recognition (NER) is an important task in NLP, which is all...
research
07/22/2021

Target-Oriented Fine-tuning for Zero-Resource Named Entity Recognition

Zero-resource named entity recognition (NER) severely suffers from data ...
research
07/24/2021

Context-aware Adversarial Training for Name Regularity Bias in Named Entity Recognition

In this work, we examine the ability of NER models to use contextual inf...
research
04/15/2021

Regularizing Models via Pointwise Mutual Information for Named Entity Recognition

In Named Entity Recognition (NER), pre-trained language models have been...
research
10/14/2022

Automatic Creation of Named Entity Recognition Datasets by Querying Phrase Representations

Most weakly supervised named entity recognition (NER) models rely on dom...

Please sign up or login with your details

Forgot password? Click here to reset