Jointly Learning to Label Sentences and Tokens

11/14/2018
by   Marek Rei, et al.
8

Learning to construct text representations in end-to-end systems can be difficult, as natural languages are highly compositional and task-specific annotated datasets are often limited in size. Methods for directly supervising language composition can allow us to guide the models based on existing knowledge, regularizing them towards more robust and interpretable representations. In this paper, we investigate how objectives at different granularities can be used to learn better language representations and we propose an architecture for jointly learning to label sentences and tokens. The predictions at each level are combined together using an attention mechanism, with token-level labels also acting as explicit supervision for composing sentence-level representations. Our experiments show that by learning to perform these tasks jointly on multiple levels, the model achieves substantial improvements for both sentence classification and sequence labeling.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/01/2020

Seeing Both the Forest and the Trees: Multi-head Attention for Joint Classification on Different Compositional Levels

In natural languages, words are used in association to construct sentenc...
research
05/06/2018

Zero-shot Sequence Labeling: Transferring Knowledge from Sentences to Tokens

Can attention- or gradient-based visualization techniques be used to inf...
research
05/24/2022

Learning for Expressive Task-Related Sentence Representations

NLP models learn sentence representations for downstream tasks by tuning...
research
09/29/2022

Perturbations and Subpopulations for Testing Robustness in Token-Based Argument Unit Recognition

Argument Unit Recognition and Classification aims at identifying argumen...
research
08/10/2018

Unsupervised Learning of Sentence Representations Using Sequence Consistency

Computing universal distributed representations of sentences is a fundam...
research
10/08/2022

Detecting Label Errors in Token Classification Data

Mislabeled examples are a common issue in real-world data, particularly ...
research
03/14/2023

Finding the Needle in a Haystack: Unsupervised Rationale Extraction from Long Text Classifiers

Long-sequence transformers are designed to improve the representation of...

Please sign up or login with your details

Forgot password? Click here to reset