A Comparative Study of Pre-trained Encoders for Low-Resource Named Entity Recognition

04/11/2022
by   Yuxuan Chen, et al.
0

Pre-trained language models (PLM) are effective components of few-shot named entity recognition (NER) approaches when augmented with continued pre-training on task-specific out-of-domain data or fine-tuning on in-domain data. However, their performance in low-resource scenarios, where such data is not available, remains an open question. We introduce an encoder evaluation framework, and use it to systematically compare the performance of state-of-the-art pre-trained representations on the task of low-resource NER. We analyze a wide range of encoders pre-trained with different strategies, model architectures, intermediate-task fine-tuning, and contrastive learning. Our experimental results across ten benchmark NER datasets in English and German show that encoder performance varies significantly, suggesting that the choice of encoder for a specific low-resource scenario needs to be carefully evaluated.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/24/2022

Formulating Few-shot Fine-tuning Towards Language Model Pre-training: A Pilot Study on Named Entity Recognition

Fine-tuning pre-trained language models has recently become a common pra...
research
12/01/2021

NER-BERT: A Pre-trained Model for Low-Resource Entity Tagging

Named entity recognition (NER) models generally perform poorly when larg...
research
11/06/2022

Prompt-based Text Entailment for Low-Resource Named Entity Recognition

Pre-trained Language Models (PLMs) have been applied in NLP tasks and ac...
research
08/31/2021

LightNER: A Lightweight Generative Framework with Prompt-guided Attention for Low-resource NER

Most existing NER methods rely on extensive labeled data for model train...
research
05/29/2023

ContrastNER: Contrastive-based Prompt Tuning for Few-shot NER

Prompt-based language models have produced encouraging results in numero...
research
05/30/2023

A Multilingual Evaluation of NER Robustness to Adversarial Inputs

Adversarial evaluations of language models typically focus on English al...
research
12/09/2022

From Clozing to Comprehending: Retrofitting Pre-trained Language Model to Pre-trained Machine Reader

We present Pre-trained Machine Reader (PMR), a novel method to retrofit ...

Please sign up or login with your details

Forgot password? Click here to reset