Chinese NER Using Lattice LSTM

05/05/2018
by   Yue Zhang, et al.
0

We investigate a lattice-structured LSTM model for Chinese NER, which encodes a sequence of input characters as well as all potential words that match a lexicon. Compared with character-based methods, our model explicitly leverages word and word sequence information. Compared with word-based methods, lattice LSTM does not suffer from segmentation errors. Gated recurrent cells allow our model to choose the most relevant characters and words from a sentence for better NER results. Experiments on various datasets show that lattice LSTM outperforms both word-based and character-based LSTM baselines, achieving the best results.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/16/2019

Simplify the Usage of Lexicon in Chinese NER

Recently, many works have tried to utilizing word lexicon to augment the...
research
10/30/2018

Subword Encoding in Lattice LSTM for Chinese Word Segmentation

We investigate a lattice LSTM network for Chinese word segmentation (CWS...
research
12/27/2017

A Gap-Based Framework for Chinese Word Segmentation via Very Deep Convolutional Networks

Most previous approaches to Chinese word segmentation can be roughly cla...
research
11/25/2019

Chinese Spelling Error Detection Using a Fusion Lattice LSTM

Spelling error detection serves as a crucial preprocessing in many natur...
research
02/18/2020

A New Clustering neural network for Chinese word segmentation

In this article I proposed a new model to achieve Chinese word segmentat...
research
05/07/2018

Sentence-State LSTM for Text Representation

Bi-directional LSTMs are a powerful tool for text representation. On the...
research
02/18/2022

TURNER: The Uncertainty-based Retrieval Framework for Chinese NER

Chinese NER is a difficult undertaking due to the ambiguity of Chinese c...

Please sign up or login with your details

Forgot password? Click here to reset