Sparse Coding of Neural Word Embeddings for Multilingual Sequence Labeling

12/21/2016
by   Gábor Berend, et al.
0

In this paper we propose and carefully evaluate a sequence labeling framework which solely utilizes sparse indicator features derived from dense distributed word representations. The proposed model obtains (near) state-of-the art performance for both part-of-speech tagging and named entity recognition for a variety of languages. Our model relies only on a few thousand sparse coding-derived features, without applying any modification of the word representations employed for the different tasks. The proposed model has favorable generalization properties as it retains over 89.8 tagging accuracy when trained at 1.2 i.e. 150 sentences per language.

READ FULL TEXT
research
11/09/2018

Neural sequence labeling for Vietnamese POS Tagging and NER

This paper presents a neural architecture for Vietnamese sequence labeli...
research
10/23/2020

Generating Adequate Distractors for Multiple-Choice Questions

This paper presents a novel approach to automatic generation of adequate...
research
10/29/2018

Learning Better Internal Structure of Words for Sequence Labeling

Character-based neural models have recently proven very useful for many ...
research
08/30/2017

An Empirical Study of Discriminative Sequence Labeling Models for Vietnamese Text Processing

This paper presents an empirical study of two widely-used sequence predi...
research
04/24/2017

Semi-supervised Multitask Learning for Sequence Labeling

We propose a sequence labeling framework with a secondary training objec...
research
07/21/2017

Optimal Hyperparameters for Deep LSTM-Networks for Sequence Labeling Tasks

Selecting optimal parameters for a neural network architecture can often...
research
03/30/2021

Locally-Contextual Nonlinear CRFs for Sequence Labeling

Linear chain conditional random fields (CRFs) combined with contextual w...

Please sign up or login with your details

Forgot password? Click here to reset