A Novel Estimator of Mutual Information for Learning to Disentangle Textual Representations

05/06/2021
by   Pierre Colombo, et al.
0

Learning disentangled representations of textual data is essential for many natural language tasks such as fair classification, style transfer and sentence generation, among others. The existent dominant approaches in the context of text data either rely on training an adversary (discriminator) that aims at making attribute values difficult to be inferred from the latent code or rely on minimising variational bounds of the mutual information between latent code and the value attribute. However, the available methods suffer of the impossibility to provide a fine-grained control of the degree (or force) of disentanglement. In contrast to adversarial methods, which are remarkably simple, although the adversary seems to be performing perfectly well during the training phase, after it is completed a fair amount of information about the undesired attribute still remains. This paper introduces a novel variational upper bound to the mutual information between an attribute and the latent code of an encoder. Our bound aims at controlling the approximation error via the Renyi's divergence, leading to both better disentangled representations and in particular, a precise control of the desirable degree of disentanglement than state-of-the-art methods proposed for textual data. Furthermore, it does not suffer from the degeneracy of other losses in multi-class scenarios. We show the superiority of this method on fair classification and on textual style transfer tasks. Additionally, we provide new insights illustrating various trade-offs in style transfer when attempting to learn disentangled representations and quality of the generated sentence.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/01/2020

Improving Disentangled Text Representation Learning with Information-Theoretic Guidance

Learning disentangled representations of natural language is essential f...
research
05/07/2022

Learning Disentangled Textual Representations via Statistical Measures of Similarity

When working with textual data, a natural application of disentangled re...
research
11/01/2018

Multiple-Attribute Text Style Transfer

The dominant approach to unsupervised "style transfer" in text is based ...
research
05/29/2019

Revision in Continuous Space: Fine-Grained Control of Text Style Transfer

Typical methods for unsupervised text style transfer often rely on two k...
research
08/04/2022

Invariant Representations with Stochastically Quantized Neural Networks

Representation learning algorithms offer the opportunity to learn invari...
research
12/20/2022

SimpleStyle: An Adaptable Style Transfer Approach

Attribute-controlled text rewriting, also known as text style-transfer, ...
research
12/16/2020

Multi-type Disentanglement without Adversarial Training

Controlling the style of natural language by disentangling the latent sp...

Please sign up or login with your details

Forgot password? Click here to reset