SEQ^3: Differentiable Sequence-to-Sequence-to-Sequence Autoencoder for Unsupervised Abstractive Sentence Compression

04/07/2019
by   Christos Baziotis, et al.
0

Neural sequence-to-sequence models are currently the dominant approach in several natural language processing tasks, but require large parallel corpora. We present a sequence-to-sequence-to-sequence autoencoder (SEQ^3), consisting of two chained encoder-decoder pairs, with words used as a sequence of discrete latent variables. We apply the proposed model to unsupervised abstractive sentence compression, where the first and last sequences are the input and reconstructed sentences, respectively, while the middle sequence is the compressed sentence. Constraining the length of the latent word sequences forces the model to distill important information from the input. A pretrained language model, acting as a prior over the latent sequences, encourages the compressed sentences to be human-readable. Continuous relaxations enable us to sample from categorical distributions, allowing gradient-based optimization, unlike alternatives that rely on reinforcement learning. The proposed model does not require parallel text-summary pairs, achieving promising results in unsupervised sentence compression on benchmark datasets.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/09/2018

Recurrent Neural Network-Based Semantic Variational Autoencoder for Sequence-to-Sequence Learning

Sequence-to-sequence (Seq2seq) models have played an import role in the ...
research
04/16/2020

Do sequence-to-sequence VAEs learn global features of sentences?

A longstanding goal in NLP is to compute global sentence representations...
research
09/07/2019

Deleter: Leveraging BERT to Perform Unsupervised Successive Text Compression

Text compression has diverse applications such as Summarization, Reading...
research
01/17/2023

Learning a Formality-Aware Japanese Sentence Representation

While the way intermediate representations are generated in encoder-deco...
research
07/01/2019

Representation, Exploration and Recommendation of Music Playlists

Playlists have become a significant part of our listening experience bec...
research
08/16/2019

An Exploratory Analysis of the Latent Structure of Process Data via Action Sequence Autoencoder

Computer simulations have become a popular tool of assessing complex ski...
research
10/30/2015

Generating Text with Deep Reinforcement Learning

We introduce a novel schema for sequence to sequence learning with a Dee...

Please sign up or login with your details

Forgot password? Click here to reset