Efficient Sequence Labeling with Actor-Critic Training

09/30/2018
by   Saeed Najafi, et al.
0

Neural approaches to sequence labeling often use a Conditional Random Field (CRF) to model their output dependencies, while Recurrent Neural Networks (RNN) are used for the same purpose in other tasks. We set out to establish RNNs as an attractive alternative to CRFs for sequence labeling. To do so, we address one of the RNN's most prominent shortcomings, the fact that it is not exposed to its own errors with the maximum-likelihood training. We frame the prediction of the output sequence as a sequential decision-making process, where we train the network with an adjusted actor-critic algorithm (AC-RNN). We comprehensively compare this strategy with maximum-likelihood training for both RNNs and CRFs on three structured-output tasks. The proposed AC-RNN efficiently matches the performance of the CRF on NER and CCG tagging, and outperforms it on Machine Transliteration. We also show that our training strategy is significantly better than other techniques for addressing RNN's exposure bias, such as Scheduled Sampling, and Self-Critical policy training.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/08/2016

Improving Recurrent Neural Networks For Sequence Labelling

In this paper we study different types of Recurrent Neural Networks (RNN...
research
03/28/2018

Actor-Critic based Training Framework for Abstractive Summarization

We present a training framework for neural abstractive summarization bas...
research
11/04/2018

Neural CRF transducers for sequence labeling

Conditional random fields (CRFs) have been shown to be one of the most s...
research
08/01/2016

Structured prediction models for RNN based sequence labeling in clinical text

Sequence labeling is a widely used method for named entity recognition a...
research
04/29/2018

From Credit Assignment to Entropy Regularization: Two New Algorithms for Neural Sequence Prediction

In this work, we study the credit assignment problem in reward augmented...
research
11/17/2015

AUC-maximized Deep Convolutional Neural Fields for Sequence Labeling

Deep Convolutional Neural Networks (DCNN) has shown excellent performanc...
research
06/14/2017

SEARNN: Training RNNs with Global-Local Losses

We propose SEARNN, a novel training algorithm for recurrent neural netwo...

Please sign up or login with your details

Forgot password? Click here to reset