Trimming and Improving Skip-thought Vectors

06/09/2017
by   Shuai Tang, et al.
0

The skip-thought model has been proven to be effective at learning sentence representations and capturing sentence semantics. In this paper, we propose a suite of techniques to trim and improve it. First, we validate a hypothesis that, given a current sentence, inferring the previous and inferring the next sentence provide similar supervision power, therefore only one decoder for predicting the next sentence is preserved in our trimmed skip-thought model. Second, we present a connection layer between encoder and decoder to help the model to generalize better on semantic relatedness tasks. Third, we found that a good word embedding initialization is also essential for learning better sentence representations. We train our model unsupervised on a large corpus with contiguous sentences, and then evaluate the trained model on 7 supervised tasks, which includes semantic relatedness, paraphrase detection, and text classification benchmarks. We empirically show that, our proposed model is a faster, lighter-weight and equally powerful alternative to the original skip-thought model.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/09/2017

Rethinking Skip-thought: A Neighborhood based Approach

We study the skip-thought model with neighborhood information as weak su...
research
06/22/2015

Skip-Thought Vectors

We describe an approach for unsupervised learning of a generic, distribu...
research
08/16/2018

Paraphrase Thought: Sentence Embedding Module Imitating Human Language Recognition

Sentence embedding is an important research topic in natural language pr...
research
04/18/2017

Representing Sentences as Low-Rank Subspaces

Sentences are important semantic units of natural language. A generic, d...
research
09/18/2018

Learning Universal Sentence Representations with Mean-Max Attention Autoencoder

In order to learn universal sentence representations, previous methods f...
research
11/19/2015

Skip-Thought Memory Networks

Question Answering (QA) is fundamental to natural language processing in...

Please sign up or login with your details

Forgot password? Click here to reset