Deep Span Representations for Named Entity Recognition

10/09/2022
by   Enwei Zhu, et al.
0

Span-based models are one of the most straightforward methods for named entity recognition (NER). Existing span-based NER systems shallowly aggregate the token representations to span representations. However, this typically results in significant ineffectiveness for long-span entities, a coupling between the representations of overlapping spans, and ultimately a performance degradation. In this study, we propose DSpERT (Deep Span Encoder Representations from Transformers), which comprises a standard Transformer and a span Transformer. The latter uses low-layered span representations as queries, and aggregates the token representations as keys and values, layer by layer from bottom to top. Thus, DSpERT produces span representations of deep semantics. With weight initialization from pretrained language models, DSpERT achieves performance higher than or competitive with recent state-of-the-art systems on eight NER benchmarks. Experimental results verify the importance of the depth for span representations, and show that DSpERT performs particularly well on long-span entities and nested structures. Further, the deep span representations are well structured and easily separable in the feature space.

READ FULL TEXT
research
04/27/2022

Propose-and-Refine: A Two-Stage Set Prediction Network for Nested Named Entity Recognition

Nested named entity recognition (nested NER) is a fundamental task in na...
research
06/09/2020

Extensive Error Analysis and a Learning-Based Evaluation of Medical Entity Recognition Systems to Approximate User Experience

When comparing entities extracted by a medical entity recognition system...
research
02/10/2023

Span-based Named Entity Recognition by Generating and Compressing Information

The information bottleneck (IB) principle has been proven effective in v...
research
05/21/2021

Boosting Span-based Joint Entity and Relation Extraction via Squence Tagging Mechanism

Span-based joint extraction simultaneously conducts named entity recogni...
research
06/11/2021

EPICURE Ensemble Pretrained Models for Extracting Cancer Mutations from Literature

To interpret the genetic profile present in a patient sample, it is nece...
research
01/02/2021

Coreference Resolution without Span Representations

Since the introduction of deep pretrained language models, most task-spe...
research
07/07/2022

Win-Win Cooperation: Bundling Sequence and Span Models for Named Entity Recognition

For Named Entity Recognition (NER), sequence labeling-based and span-bas...

Please sign up or login with your details

Forgot password? Click here to reset