Vec2Sent: Probing Sentence Embeddings with Natural Language Generation

11/01/2020
by   Martin Kerscher, et al.
0

We introspect black-box sentence embeddings by conditionally generating from them with the objective to retrieve the underlying discrete sentence. We perceive of this as a new unsupervised probing task and show that it correlates well with downstream task performance. We also illustrate how the language generated from different encoders differs. We apply our approach to generate sentence analogies from sentence embeddings.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/03/2018

What you can cram into a single vector: Probing sentence embeddings for linguistic properties

Although much effort has recently been devoted to training high-quality ...
research
08/24/2023

Multi-BERT for Embeddings for Recommendation System

In this paper, we propose a novel approach for generating document embed...
research
09/06/2019

Natural Adversarial Sentence Generation with Gradient-based Perturbation

This work proposes a novel algorithm to generate natural language advers...
research
04/05/2021

WhiteningBERT: An Easy Unsupervised Sentence Embedding Approach

Producing the embedding of a sentence in an unsupervised way is valuable...
research
11/15/2017

Pushing the Limits of Paraphrastic Sentence Embeddings with Millions of Machine Translations

We extend the work of Wieting et al. (2017), back-translating a large pa...
research
04/03/2019

The Effect of Downstream Classification Tasks for Evaluating Sentence Embeddings

One popular method for quantitatively evaluating the performance of sent...
research
10/04/2019

Neural Language Priors

The choice of sentence encoder architecture reflects assumptions about h...

Please sign up or login with your details

Forgot password? Click here to reset