Recurrent Neural Network-Based Semantic Variational Autoencoder for Sequence-to-Sequence Learning

02/09/2018
by   Myeongjun Jang, et al.
0

Sequence-to-sequence (Seq2seq) models have played an import role in the recent success of various natural language processing methods, such as machine translation, text summarization, and speech recognition. However, current Seq2seq models have trouble preserving global latent information from a long sequence of words. Variational autoencoder (VAE) alleviates this problem by learning a continuous semantic space of the input sentence. However, it does not solve the problem completely. In this paper, we propose a new recurrent neural network (RNN)-based Seq2seq model, RNN semantic variational autoencoder (RNN--SVAE), to better capture the global latent information of a sequence of words. To consider the words in a sentence equally, without regard to its position within the sentence, we construct a document information vector using the attention information between the final state of the encoder and every prior hidden state. Then, we combine this document information vector with the final hidden state of the bi-directional RNN encoder to construct the global latent vector, which becomes the output of the encoder part. Then, the mean and standard deviation of the continuous semantic space are learned to take advantage of the variational method. Experimental results of three natural language tasks (i.e., language modeling, missing word imputation, paraphrase identification) confirm that the proposed RNN--SVAE yields higher performance than two benchmark models.

READ FULL TEXT
research
11/19/2015

Generating Sentences from a Continuous Space

The standard recurrent neural network language model (RNNLM) generates s...
research
04/07/2019

SEQ^3: Differentiable Sequence-to-Sequence-to-Sequence Autoencoder for Unsupervised Abstractive Sentence Compression

Neural sequence-to-sequence models are currently the dominant approach i...
research
11/13/2018

Modeling Local Dependence in Natural Language with Multi-channel Recurrent Neural Networks

Recurrent Neural Networks (RNNs) have been widely used in processing nat...
research
06/05/2018

Explaining Away Syntactic Structure in Semantic Document Representations

Most generative document models act on bag-of-words input in an attempt ...
research
04/16/2020

Do sequence-to-sequence VAEs learn global features of sentences?

A longstanding goal in NLP is to compute global sentence representations...
research
03/13/2023

Learning Transductions and Alignments with RNN Seq2seq Models

The paper studies the capabilities of Recurrent-Neural-Network sequence ...
research
06/09/2018

Method to Annotate Arrhythmias by Deep Network

This study targets to automatically annotate on arrhythmia by deep netwo...

Please sign up or login with your details

Forgot password? Click here to reset