Learning Better Internal Structure of Words for Sequence Labeling

10/29/2018
by   Yingwei Xin, et al.
0

Character-based neural models have recently proven very useful for many NLP tasks. However, there is a gap of sophistication between methods for learning representations of sentences and words. While most character models for learning representations of sentences are deep and complex, models for learning representations of words are shallow and simple. Also, in spite of considerable research on learning character embeddings, it is still not clear which kind of architecture is the best for capturing character-to-word representations. To address these questions, we first investigate the gaps between methods for learning word and sentence representations. We conduct detailed experiments and comparisons of different state-of-the-art convolutional models, and also investigate the advantages and disadvantages of their constituents. Furthermore, we propose IntNet, a funnel-shaped wide convolutional neural architecture with no down-sampling for learning representations of the internal structure of words by composing their characters from limited, supervised training corpora. We evaluate our proposed model on six sequence labeling datasets, including named entity recognition, part-of-speech tagging, and syntactic chunking. Our in-depth analysis shows that IntNet significantly outperforms other character embedding models and obtains new state-of-the-art performance without relying on any external knowledge or resources.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/10/2016

Charagram: Embedding Words and Sentences via Character n-grams

We present Charagram embeddings, a simple approach for learning characte...
research
11/14/2016

Attending to Characters in Neural Sequence Labeling Models

Sequence labeling architectures use word embeddings for capturing simila...
research
12/21/2016

Sparse Coding of Neural Word Embeddings for Multilingual Sequence Labeling

In this paper we propose and carefully evaluate a sequence labeling fram...
research
09/26/2018

Deep contextualized word representations for detecting sarcasm and irony

Predicting context-dependent and non-literal utterances like sarcastic a...
research
05/29/2019

Learning Task-specific Representation for Novel Words in Sequence Labeling

Word representation is a key component in neural-network-based sequence ...
research
02/10/2016

Learning Distributed Representations of Sentences from Unlabelled Data

Unsupervised methods for learning distributed representations of words a...
research
08/03/2018

Efficient Purely Convolutional Text Encoding

In this work, we focus on a lightweight convolutional architecture that ...

Please sign up or login with your details

Forgot password? Click here to reset