Discrete Autoencoders for Sequence Models

01/29/2018
by   Łukasz Kaiser, et al.
0

Recurrent models for sequences have been recently successful at many tasks, especially for language modeling and machine translation. Nevertheless, it remains challenging to extract good representations from these models. For instance, even though language has a clear hierarchical structure going from characters through words to sentences, it is not apparent in current language models. We propose to improve the representation in sequence models by augmenting current approaches with an autoencoder that is forced to compress the sequence through an intermediate discrete latent space. In order to propagate gradients though this discrete representation we introduce an improved semantic hashing technique. We show that this technique performs well on a newly proposed quantitative efficiency measure. We also analyze latent codes produced by the model showing how they correspond to words and phrases. Finally, we present an application of the autoencoder-augmented model to generating diverse translations.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/22/2020

Discretized Bottleneck in VAE: Posterior-Collapse-Free Sequence-to-Sequence Learning

Variational autoencoders (VAEs) are important tools in end-to-end repres...
research
07/13/2022

N-Grammer: Augmenting Transformers with latent n-grams

Transformer models have recently emerged as one of the foundational mode...
research
11/19/2015

Generating Sentences from a Continuous Space

The standard recurrent neural network language model (RNNLM) generates s...
research
12/01/2020

Latent Programmer: Discrete Latent Codes for Program Synthesis

In many sequence learning tasks, such as program synthesis and document ...
research
05/18/2021

Representation Learning in Sequence to Sequence Tasks: Multi-filter Gaussian Mixture Autoencoder

Heterogeneity of sentences exists in sequence to sequence tasks such as ...
research
08/28/2018

Hierarchical Quantized Representations for Script Generation

Scripts define knowledge about how everyday scenarios (such as going to ...
research
07/08/2022

Hidden Schema Networks

Most modern language models infer representations that, albeit powerful,...

Please sign up or login with your details

Forgot password? Click here to reset