Why Attention? Analyzing and Remedying BiLSTM Deficiency in Modeling Cross-Context for NER

10/07/2019
by   Peng-Hsuan Li, et al.
0

State-of-the-art approaches of NER have used sequence-labeling BiLSTM as a core module. This paper formally shows the limitation of BiLSTM in modeling cross-context patterns. Two types of simple cross-structures – self-attention and Cross-BiLSTM – are shown to effectively remedy the problem. On both OntoNotes 5.0 and WNUT 2017, clear and consistent improvements are achieved over bare-bone models, up to 8.7 analyses across several aspects of the improvements, especially the identification of multi-token mentions, are further given.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/29/2019

Remedying BiLSTM-CNN Deficiency in Modeling Cross-Context for NER

Recent researches prevalently used BiLSTM-CNN as a core module for NER i...
research
06/02/2020

Exploring Cross-sentence Contexts for Named Entity Recognition with BERT

Named entity recognition (NER) is frequently addressed as a sequence cla...
research
04/19/2022

On the Locality of Attention in Direct Speech Translation

Transformers have achieved state-of-the-art results across multiple NLP ...
research
06/06/2019

GCDT: A Global Context Enhanced Deep Transition Architecture for Sequence Labeling

Current state-of-the-art systems for sequence labeling are typically bas...
research
03/26/2018

Self-Attentional Acoustic Models

Self-attention is a method of encoding sequences of vectors by relating ...
research
05/30/2023

Comparing and combining some popular NER approaches on Biomedical tasks

We compare three simple and popular approaches for NER: 1) SEQ (sequence...
research
09/01/2021

Joint Graph Learning and Matching for Semantic Feature Correspondence

In recent years, powered by the learned discriminative representation vi...

Please sign up or login with your details

Forgot password? Click here to reset