Disentangled Recurrent Wasserstein Autoencoder

01/19/2021
by   Jun Han, et al.
12

Learning disentangled representations leads to interpretable models and facilitates data generation with style transfer, which has been extensively studied on static data such as images in an unsupervised learning framework. However, only a few works have explored unsupervised disentangled sequential representation learning due to challenges of generating sequential data. In this paper, we propose recurrent Wasserstein Autoencoder (R-WAE), a new framework for generative modeling of sequential data. R-WAE disentangles the representation of an input sequence into static and dynamic factors (i.e., time-invariant and time-varying parts). Our theoretical analysis shows that, R-WAE minimizes an upper bound of a penalized form of the Wasserstein distance between model distribution and sequential data distribution, and simultaneously maximizes the mutual information between input data and different disentangled latent factors, respectively. This is superior to (recurrent) VAE which does not explicitly enforce mutual information maximization between input data and disentangled latent representations. When the number of actions in sequential data is available as weak supervision information, R-WAE is extended to learn a categorical latent representation of actions to improve its disentanglement. Experiments on a variety of datasets show that our models outperform other baselines with the same settings in terms of disentanglement and unconditional video generation both quantitatively and qualitatively.

READ FULL TEXT

page 6

page 8

page 16

page 17

page 18

page 20

page 21

research
10/22/2021

Contrastively Disentangled Sequential Variational Autoencoder

Self-supervised disentangled representation learning is a critical task ...
research
02/22/2019

FAVAE: Sequence Disentanglement using Information Bottleneck Principle

We propose the factorized action variational autoencoder (FAVAE), a stat...
research
05/23/2020

S3VAE: Self-Supervised Sequential VAE for Representation Disentanglement and Data Generation

We propose a sequential variational autoencoder to learn disentangled re...
research
04/18/2018

Deep Generative Networks For Sequence Prediction

This thesis investigates unsupervised time series representation learnin...
research
03/30/2023

Multifactor Sequential Disentanglement via Structured Koopman Autoencoders

Disentangling complex data to its latent factors of variation is a funda...
research
06/07/2019

Disentangled State Space Representations

Sequential data often originates from diverse domains across which stati...
research
01/08/2020

Disentangling Representations using Gaussian Processes in Variational Autoencoders for Video Prediction

We introduce MGP-VAE, a variational autoencoder which uses Gaussian proc...

Please sign up or login with your details

Forgot password? Click here to reset