Porous Lattice-based Transformer Encoder for Chinese NER

11/07/2019
by   Xue Mengge, et al.
0

Incorporating lattices into character-level Chinese named entity recognition is an effective method to exploit explicit word information. Recent works extend recurrent and convolutional neural networks to model lattice inputs. However, due to the DAG structure or the variable-sized potential word set for lattice inputs, these models prevent the convenient use of batched computation, resulting in serious inefficient. In this paper, we propose a porous lattice-based transformer encoder for Chinese named entity recognition, which is capable to better exploit the GPU parallelism and batch the computation owing to the mask mechanism in transformer. We first investigate the lattice-aware self-attention coupled with relative position representations to explore effective word information in the lattice structure. Besides, to strengthen the local dependencies among neighboring tokens, we propose a novel porous structure during self-attentional computation processing, in which every two non-neighboring tokens are connected through a shared pivot node. Experimental results on four datasets show that our model performs up to 9.47 times faster than state-of-the-art models, while is roughly on a par with its performance. The source code of this paper can be obtained from https://github.com/xxx/xxx.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/24/2020

FLAT: Chinese NER Using Flat-Lattice Transformer

Recently, the character-word lattice structure has been proved to be eff...
research
07/12/2021

MECT: Multi-Metadata Embedding based Cross-Transformer for Chinese Named Entity Recognition

Recently, word enhancement has become very popular for Chinese Named Ent...
research
05/12/2022

NFLAT: Non-Flat-Lattice Transformer for Chinese Named Entity Recognition

Recently, Flat-LAttice Transformer (FLAT) has achieved great success in ...
research
11/10/2019

TENER: Adapting Transformer Encoder for Name Entity Recognition

The Bidirectional long short-term memory networks (BiLSTM) have been wid...
research
08/16/2019

Simplify the Usage of Lexicon in Chinese NER

Recently, many works have tried to utilizing word lexicon to augment the...
research
12/17/2019

Chinese Named Entity Recognition Augmented with Lexicon Memory

Inspired by a concept of content-addressable retrieval from cognitive sc...
research
06/04/2019

Self-Attentional Models for Lattice Inputs

Lattices are an efficient and effective method to encode ambiguity of up...

Please sign up or login with your details

Forgot password? Click here to reset