Locally-Contextual Nonlinear CRFs for Sequence Labeling

03/30/2021
by   Harshil Shah, et al.
0

Linear chain conditional random fields (CRFs) combined with contextual word embeddings have achieved state of the art performance on sequence labeling tasks. In many of these tasks, the identity of the neighboring words is often the most useful contextual information when predicting the label of a given word. However, contextual embeddings are usually trained in a task-agnostic manner. This means that although they may encode information about the neighboring words, it is not guaranteed. It can therefore be beneficial to design the sequence labeling architecture to directly extract this information from the embeddings. We propose locally-contextual nonlinear CRFs for sequence labeling. Our approach directly incorporates information from the neighboring embeddings when predicting the label for a given word, and parametrizes the potential functions using deep neural networks. Our model serves as a drop-in replacement for the linear chain CRF, consistently outperforming it in our ablation study. On a variety of tasks, our results are competitive with those of the best published methods. In particular, we outperform the previous state of the art on chunking on CoNLL 2000 and named entity recognition on OntoNotes 5.0 English.

READ FULL TEXT
research
11/11/2020

An Investigation of Potential Function Designs for Neural CRF

The neural linear-chain CRF model is one of the most widely-used approac...
research
06/05/2020

Sentence Compression as Deletion with Contextual Embeddings

Sentence compression is the task of creating a shorter version of an inp...
research
05/10/2018

Hybrid semi-Markov CRF for Neural Sequence Labeling

This paper proposes hybrid semi-Markov conditional random fields (SCRFs)...
research
08/30/2017

An Empirical Study of Discriminative Sequence Labeling Models for Vietnamese Text Processing

This paper presents an empirical study of two widely-used sequence predi...
research
08/22/2023

Generalising sequence models for epigenome predictions with tissue and assay embeddings

Sequence modelling approaches for epigenetic profile prediction have rec...
research
12/21/2016

Sparse Coding of Neural Word Embeddings for Multilingual Sequence Labeling

In this paper we propose and carefully evaluate a sequence labeling fram...
research
09/06/2020

MIDAS at SemEval-2020 Task 10: Emphasis Selection using Label Distribution Learning and Contextual Embeddings

This paper presents our submission to the SemEval 2020 - Task 10 on emph...

Please sign up or login with your details

Forgot password? Click here to reset