GCDT: A Global Context Enhanced Deep Transition Architecture for Sequence Labeling

06/06/2019
by   Yijin Liu, et al.
0

Current state-of-the-art systems for sequence labeling are typically based on the family of Recurrent Neural Networks (RNNs). However, the shallow connections between consecutive hidden states of RNNs and insufficient modeling of global information restrict the potential performance of those models. In this paper, we try to address these issues, and thus propose a Global Context enhanced Deep Transition architecture for sequence labeling named GCDT. We deepen the state transition path at each position in a sentence, and further assign every token with a global representation learned from the entire sentence. Experiments on two standard sequence labeling tasks show that, given only training data and the ubiquitous word embeddings (Glove), our GCDT achieves 91.96 F1 on the CoNLL03 NER task and 95.43 F1 on the CoNLL2000 Chunking task, which outperforms the best reported results under the same settings. Furthermore, by leveraging BERT as an additional resource, we establish new state-of-the-art results with 93.47 F1 on NER and 97.30 F1 on Chunking.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/06/2019

Hierarchical Contextualized Representation for Named Entity Recognition

Current named entity recognition (NER) models are typically based on the...
research
05/31/2023

A Global Context Mechanism for Sequence Labeling

Sequential labeling tasks necessitate the computation of sentence repres...
research
05/30/2023

Comparing and combining some popular NER approaches on Biomedical tasks

We compare three simple and popular approaches for NER: 1) SEQ (sequence...
research
10/07/2020

OpenIE6: Iterative Grid Labeling and Coordination Analysis for Open Information Extraction

A recent state-of-the-art neural open information extraction (OpenIE) sy...
research
10/07/2019

Why Attention? Analyzing and Remedying BiLSTM Deficiency in Modeling Cross-Context for NER

State-of-the-art approaches of NER have used sequence-labeling BiLSTM as...
research
10/23/2020

NLNDE at CANTEMIST: Neural Sequence Labeling and Parsing Approaches for Clinical Concept Extraction

The recognition and normalization of clinical information, such as tumor...
research
06/04/2019

Toward Grammatical Error Detection from Sentence Labels: Zero-shot Sequence Labeling with CNNs and Contextualized Embeddings

Zero-shot grammatical error detection is the task of tagging token-level...

Please sign up or login with your details

Forgot password? Click here to reset