Deconvolutional Paragraph Representation Learning

08/16/2017
by   Yizhe Zhang, et al.
0

Learning latent representations from long text sequences is an important first step in many natural language processing applications. Recurrent Neural Networks (RNNs) have become a cornerstone for this challenging task. However, the quality of sentences during RNN-based decoding (reconstruction) decreases with the length of the text. We propose a sequence-to-sequence, purely convolutional and deconvolutional autoencoding framework that is free of the above issue, while also being computationally efficient. The proposed method is simple, easy to implement and can be leveraged as a building block for many applications. We show empirically that compared to RNNs, our framework is better at reconstructing and correcting long paragraphs. Quantitative evaluation on semi-supervised text classification and summarization tasks demonstrate the potential for better utilization of long unlabeled text data.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/04/2015

Semi-supervised Sequence Learning

We present two approaches that use unlabeled data to improve sequence le...
research
05/30/2019

Particle Filter Recurrent Neural Networks

Recurrent neural networks (RNNs) have been extraordinarily successful fo...
research
11/05/2018

Structured Neural Summarization

Summarization of long sequences into a concise statement is a core probl...
research
09/28/2018

Learning Robust, Transferable Sentence Representations for Text Classification

Despite deep recurrent neural networks (RNNs) demonstrate strong perform...
research
10/28/2020

The geometry of integration in text classification RNNs

Despite the widespread application of recurrent neural networks (RNNs) a...
research
05/16/2018

Learning to Write with Cooperative Discriminators

Recurrent Neural Networks (RNNs) are powerful autoregressive sequence mo...
research
05/15/2018

SoPa: Bridging CNNs, RNNs, and Weighted Finite-State Machines

Recurrent and convolutional neural networks comprise two distinct famili...

Please sign up or login with your details

Forgot password? Click here to reset