Does Higher Order LSTM Have Better Accuracy in Chunking and Named Entity Recognition?

11/22/2017
by   Yi Zhang, et al.
0

Current researches usually employ single order setting by default when dealing with sequence labeling tasks. In our work, "order" means the number of tags that a prediction involves at every time step. High order models tend to capture more dependency information among tags. We first propose a simple method that low order models can be easily extended to high order models. To our surprise, the high order models which are supposed to capture more dependency information behave worse when increasing the order. We suppose that forcing neural networks to learn complex structure may lead to overfitting. To deal with the problem, we propose a method which combines low order and high order information together to decode the tag sequence. The proposed method, multi-order decoding (MOD), keeps the scalability to high order models with a pruning technique. MOD achieves higher accuracies than existing methods of single order setting. It results in a 21 in chunking and an error reduction over 23 available at https://github.com/lancopku/Multi-Order-Decoding.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/06/2020

An Exploration of Arbitrary-Order Sequence Labeling via Energy-Based Inference Networks

Many tasks in natural language processing involve predicting structured ...
research
11/25/2017

Complex Structure Leads to Overfitting: A Structure Regularization Decoding Method for Natural Language Processing

Recent systems on structured prediction focus on increasing the level of...
research
07/16/2020

Dual Graph Embedding for Object-Tag LinkPrediction on the Knowledge Graph

Knowledge graphs (KGs) composed of users, objects, and tags are widely u...
research
06/18/2023

Transferring Neural Potentials For High Order Dependency Parsing

High order dependency parsing leverages high order features such as sibl...
research
11/11/2022

Getting the Most out of Simile Recognition

Simile recognition involves two subtasks: simile sentence classification...
research
07/17/2023

HOPE: High-order Polynomial Expansion of Black-box Neural Networks

Despite their remarkable performance, deep neural networks remain mostly...
research
08/10/2023

Information decomposition reveals hidden high-order contributions to temporal irreversibility

Temporal irreversibility, often referred to as the arrow of time, is a f...

Please sign up or login with your details

Forgot password? Click here to reset