AriEL: volume coding for sentence generation

03/30/2020
by   Luca Celotti, et al.
0

Mapping sequences of discrete data to a point in a continuous space makes it difficult to retrieve those sequences via random sampling. Mapping the input to a volume would make it easier to retrieve at test time, and that's the strategy followed by the family of approaches based on Variational Autoencoder. However the fact that they are at the same time optimizing for prediction and for smoothness of representation, forces them to trade-off between the two. We improve on the performance of some of the standard methods in deep learning to generate sentences by uniformly sampling a continuous space. We do it by proposing AriEL, that constructs volumes in a continuous space, without the need of encouraging the creation of volumes through the loss function. We first benchmark on a toy grammar, that allows to automatically evaluate the language learned and generated by the models. Then, we benchmark on a real dataset of human dialogues. Our results indicate that the random access to the stored information is dramatically improved, and our method AriEL is able to generate a wider variety of correct language by randomly sampling the latent space. VAE follows in performance for the toy dataset while, AE and Transformer follow for the real dataset. This partially supports to the hypothesis that encoding information into volumes instead of into points, can lead to improved retrieval of learned information with random sampling. This can lead to better generators and we also discuss potential disadvantages.

READ FULL TEXT
research
09/15/2022

A Geometric Perspective on Variational Autoencoders

This paper introduces a new interpretation of the Variational Autoencode...
research
03/29/2023

SC-VAE: Sparse Coding-based Variational Autoencoder

Learning rich data representations from unlabeled data is a key challeng...
research
09/23/2022

GLSO: Grammar-guided Latent Space Optimization for Sample-efficient Robot Design Automation

Robots have been used in all sorts of automation, and yet the design of ...
research
03/06/2017

Grammar Variational Autoencoder

Deep generative models have been wildly successful at learning coherent ...
research
08/27/2018

Natural Language Generation with Neural Variational Models

In this thesis, we explore the use of deep neural networks for generatio...
research
04/09/2020

Exemplar VAEs for Exemplar based Generation and Data Augmentation

This paper presents a framework for exemplar based generative modeling, ...

Please sign up or login with your details

Forgot password? Click here to reset