Encouraging Paragraph Embeddings to Remember Sentence Identity Improves Classification

06/09/2019
by   Tu Vu, et al.
0

While paragraph embedding models are remarkably effective for downstream classification tasks, what they learn and encode into a single vector remains opaque. In this paper, we investigate a state-of-the-art paragraph embedding method proposed by Zhang et al. (2017) and discover that it cannot reliably tell whether a given sentence occurs in the input paragraph or not. We formulate a sentence content task to probe for this basic linguistic property and find that even a much simpler bag-of-words method has no trouble solving it. This result motivates us to replace the reconstruction-based objective of Zhang et al. (2017) with our sentence content probe objective in a semi-supervised setting. Despite its simplicity, our objective improves over paragraph reconstruction in terms of (1) downstream classification accuracies on benchmark datasets, (2) faster training, and (3) better generalization ability.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/11/2021

Improved LCAs for constructing spanners

In this paper we study the problem of constructing spanners in a local m...
research
04/16/2021

Are Classes Clusters?

Sentence embedding models aim to provide general purpose embeddings for ...
research
04/03/2019

The Effect of Downstream Classification Tasks for Evaluating Sentence Embeddings

One popular method for quantitatively evaluating the performance of sent...
research
05/15/2019

What do you learn from context? Probing for sentence structure in contextualized word representations

Contextualized representation models such as ELMo (Peters et al., 2018a)...
research
06/16/2018

Evaluation of sentence embeddings in downstream and linguistic probing tasks

Despite the fast developmental pace of new sentence embedding methods, i...
research
09/15/2020

An information theoretic view on selecting linguistic probes

There is increasing interest in assessing the linguistic knowledge encod...
research
05/24/2019

Human vs. Muppet: A Conservative Estimate of HumanPerformance on the GLUE Benchmark

The GLUE benchmark (Wang et al., 2019b) is a suite of language understan...

Please sign up or login with your details

Forgot password? Click here to reset