Learning Universal Sentence Representations with Mean-Max Attention Autoencoder

09/18/2018
by   Minghua Zhang, et al.
0

In order to learn universal sentence representations, previous methods focus on complex recurrent neural networks or supervised learning. In this paper, we propose a mean-max attention autoencoder (mean-max AAE) within the encoder-decoder framework. Our autoencoder rely entirely on the MultiHead self-attention mechanism to reconstruct the input sequence. In the encoding we propose a mean-max strategy that applies both mean and max pooling operations over the hidden vectors to capture diverse information of the input. To enable the information to steer the reconstruction process dynamically, the decoder performs attention over the mean-max representation. By training our model on a large collection of unlabelled data, we obtain high-quality representations of sentences. Experimental results on a broad range of 10 transfer tasks demonstrate that our model outperforms the state-of-the-art unsupervised single methods, including the classical skip-thoughts and the advanced skip-thoughts+LN model. Furthermore, compared with the traditional recurrent neural network, our mean-max AAE greatly reduce the training time.

READ FULL TEXT
research
11/23/2016

Learning Generic Sentence Representations Using Convolutional Neural Networks

We propose a new encoder-decoder approach to learn distributed sentence ...
research
09/14/2016

Neural Machine Transliteration: Preliminary Results

Machine transliteration is the process of automatically transforming the...
research
08/27/2018

Natural Language Inference with Hierarchical BiLSTM Max Pooling Architecture

Recurrent neural networks have proven to be very effective for natural l...
research
06/09/2017

Trimming and Improving Skip-thought Vectors

The skip-thought model has been proven to be effective at learning sente...
research
08/25/2023

MMBAttn: Max-Mean and Bit-wise Attention for CTR Prediction

With the increasing complexity and scale of click-through rate (CTR) pre...
research
09/15/2020

Recurrent autoencoder with sequence-aware encoding

Recurrent Neural Networks (RNN) received a vast amount of attention last...
research
08/27/2013

Improving Sparse Associative Memories by Escaping from Bogus Fixed Points

The Gripon-Berrou neural network (GBNN) is a recently invented recurrent...

Please sign up or login with your details

Forgot password? Click here to reset