Disentangling Generative Factors in Natural Language with Discrete Variational Autoencoders

09/15/2021
by   Giangiacomo Mercatali, et al.
0

The ability of learning disentangled representations represents a major step for interpretable NLP systems as it allows latent linguistic features to be controlled. Most approaches to disentanglement rely on continuous variables, both for images and text. We argue that despite being suitable for image datasets, continuous variables may not be ideal to model features of textual data, due to the fact that most generative factors in text are discrete. We propose a Variational Autoencoder based method which models language features as discrete variables and encourages independence between variables for learning disentangled representations. The proposed model outperforms continuous and discrete baselines on several qualitative and quantitative benchmarks for disentanglement as well as on a text style transfer downstream application.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/31/2018

Joint-VAE: Learning Disentangled Joint Continuous and Discrete Representations

We present a framework for learning disentangled and interpretable joint...
research
11/24/2017

Quantifying the Effects of Enforcing Disentanglement on Variational Autoencoders

The notion of disentangled autoencoders was proposed as an extension to ...
research
07/26/2023

Learning Disentangled Discrete Representations

Recent successes in image generation, model-based reinforcement learning...
research
06/01/2020

Improving Disentangled Text Representation Learning with Information-Theoretic Guidance

Learning disentangled representations of natural language is essential f...
research
08/20/2021

Extracting Qualitative Causal Structure with Transformer-Based NLP

Qualitative causal relationships compactly express the direction, depend...
research
04/28/2020

Learning Interpretable and Discrete Representations with Adversarial Training for Unsupervised Text Classification

Learning continuous representations from unlabeled textual data has been...
research
05/04/2017

KATE: K-Competitive Autoencoder for Text

Autoencoders have been successful in learning meaningful representations...

Please sign up or login with your details

Forgot password? Click here to reset