Revisiting Recurrent Networks for Paraphrastic Sentence Embeddings

04/30/2017
by   John Wieting, et al.
0

We consider the problem of learning general-purpose, paraphrastic sentence embeddings, revisiting the setting of Wieting et al. (2016b). While they found LSTM recurrent networks to underperform word averaging, we present several developments that together produce the opposite conclusion. These include training on sentence pairs rather than phrase pairs, averaging states to represent sequences, and regularizing aggressively. These improve LSTMs in both transfer learning and supervised settings. We also introduce a new recurrent architecture, the Gated Recurrent Averaging Network, that is inspired by averaging and LSTMs while outperforming them both. We analyze our learned models, finding evidence of preferences for particular parts of speech and dependency relations.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/25/2015

Towards Universal Paraphrastic Sentence Embeddings

We consider the problem of learning general-purpose, paraphrastic senten...
research
06/06/2017

Learning Paraphrastic Sentence Embeddings from Back-Translated Bitext

We consider the problem of learning general-purpose, paraphrastic senten...
research
05/22/2018

Learning sentence embeddings using Recursive Networks

Learning sentence vectors that generalise well is a challenging task. In...
research
09/21/2018

Paraphrase Detection on Noisy Subtitles in Six Languages

We perform automatic paraphrase detection on subtitle data from the Opus...
research
11/03/2017

Fine-tuning Tree-LSTM for phrase-level sentiment classification on a Polish dependency treebank. Submission to PolEval task 2

We describe a variant of Child-Sum Tree-LSTM deep neural network (Tai et...
research
11/16/2016

A Way out of the Odyssey: Analyzing and Combining Recent Insights for LSTMs

LSTMs have become a basic building block for many deep NLP models. In re...
research
02/12/2019

PAC-Bayes Analysis of Sentence Representation

Learning sentence vectors from an unlabeled corpus has attracted attenti...

Please sign up or login with your details

Forgot password? Click here to reset