The Transitive Information Theory and its Application to Deep Generative Models

03/09/2022
by   Trung Ngo, et al.
0

Paradoxically, a Variational Autoencoder (VAE) could be pushed in two opposite directions, utilizing powerful decoder model for generating realistic images but collapsing the learned representation, or increasing regularization coefficient for disentangling representation but ultimately generating blurry examples. Existing methods narrow the issues to the rate-distortion trade-off between compression and reconstruction. We argue that a good reconstruction model does learn high capacity latents that encode more details, however, its use is hindered by two major issues: the prior is random noise which is completely detached from the posterior and allow no controllability in the generation; mean-field variational inference doesn't enforce hierarchy structure which makes the task of recombining those units into plausible novel output infeasible. As a result, we develop a system that learns a hierarchy of disentangled representation together with a mechanism for recombining the learned representation for generalization. This is achieved by introducing a minimal amount of inductive bias to learn controllable prior for the VAE. The idea is supported by here developed transitive information theory, that is, the mutual information between two target variables could alternately be maximized through the mutual information to the third variable, thus bypassing the rate-distortion bottleneck in VAE design. In particular, we show that our model, named SemafoVAE (inspired by the similar concept in computer science), could generate high-quality examples in a controllable manner, perform smooth traversals of the disentangled factors and intervention at a different level of representation hierarchy.

READ FULL TEXT

page 9

page 14

page 15

page 24

page 25

page 26

page 27

research
09/06/2018

Discovering Influential Factors in Variational Autoencoder

In the field of machine learning, it is still a critical issue to identi...
research
04/13/2020

Controllable Variational Autoencoder

Variational Autoencoders (VAE) and their variants have been widely used ...
research
05/23/2023

Disentangled Variational Autoencoder for Emotion Recognition in Conversations

In Emotion Recognition in Conversations (ERC), the emotions of target ut...
research
12/02/2019

Information bottleneck through variational glasses

Information bottleneck (IB) principle [1] has become an important elemen...
research
05/27/2019

Wyner VAE: Joint and Conditional Generation with Succinct Common Representation Learning

A new variational autoencoder (VAE) model is proposed that learns a succ...
research
05/25/2019

The variational infomax autoencoder

We propose the Variational InfoMax AutoEncoder (VIMAE), a method to trai...
research
09/14/2023

Dataset Size Dependence of Rate-Distortion Curve and Threshold of Posterior Collapse in Linear VAE

In the Variational Autoencoder (VAE), the variational posterior often al...

Please sign up or login with your details

Forgot password? Click here to reset