NAT: Noise-Aware Training for Robust Neural Sequence Labeling

05/14/2020
by   Marcin Namysl, et al.
0

Sequence labeling systems should perform reliably not only under ideal conditions but also with corrupted inputs - as these systems often process user-generated text or follow an error-prone upstream component. To this end, we formulate the noisy sequence labeling problem, where the input may undergo an unknown noising process and propose two Noise-Aware Training (NAT) objectives that improve robustness of sequence labeling performed on perturbed input: Our data augmentation method trains a neural model using a mixture of clean and noisy samples, whereas our stability training algorithm encourages the model to create a noise-invariant latent representation. We employ a vanilla noise model at training time. For evaluation, we use both the original data and its variants perturbed with real OCR errors and misspellings. Extensive experiments on English and German named entity recognition benchmarks confirmed that NAT consistently improved robustness of popular sequence labeling models, preserving accuracy on the original input. We make our code and data publicly available for the research community.

READ FULL TEXT

page 7

page 8

page 17

research
05/25/2021

Empirical Error Modeling Improves Robustness of Noisy Neural Sequence Labeling

Despite recent advances, standard sequence labeling systems often fail w...
research
02/17/2023

Uncertainty-aware Self-training for Low-resource Neural Sequence Labeling

Neural sequence labeling (NSL) aims at assigning labels for input langua...
research
11/13/2019

Robustness to Capitalization Errors in Named Entity Recognition

Robustness to capitalization errors is a highly desirable characteristic...
research
10/05/2020

SeqMix: Augmenting Active Sequence Labeling via Sequence Mixup

Active learning is an important technique for low-resource sequence labe...
research
05/29/2019

Learning Task-specific Representation for Novel Words in Sequence Labeling

Word representation is a key component in neural-network-based sequence ...
research
10/21/2020

German's Next Language Model

In this work we present the experiments which lead to the creation of ou...
research
10/18/2022

Denoising Enhanced Distantly Supervised Ultrafine Entity Typing

Recently, the task of distantly supervised (DS) ultra-fine entity typing...

Please sign up or login with your details

Forgot password? Click here to reset