Explicitly disentangling image content from translation and rotation with spatial-VAE

09/25/2019
by   Tristan Bepler, et al.
7

Given an image dataset, we are often interested in finding data generative factors that encode semantic content independently from pose variables such as rotation and translation. However, current disentanglement approaches do not impose any specific structure on the learned latent representations. We propose a method for explicitly disentangling image rotation and translation from other unstructured latent factors in a variational autoencoder (VAE) framework. By formulating the generative model as a function of the spatial coordinate, we make the reconstruction error differentiable with respect to latent translation and rotation parameters. This formulation allows us to train a neural network to perform approximate inference on these latent variables while explicitly constraining them to only represent rotation and translation. We demonstrate that this framework, termed spatial-VAE, effectively learns latent representations that disentangle image rotation and translation from content and improves reconstruction over standard VAEs on several benchmark datasets, including applications to modeling continuous 2-D views of proteins from single particle electron microscopy and galaxies in astronomical images.

READ FULL TEXT

page 8

page 14

page 16

page 17

page 18

research
10/24/2022

Unsupervised Object Representation Learning using Translation and Rotation Group Equivariant VAE

In many imaging modalities, objects of interest can occur in a variety o...
research
03/22/2022

Upmixing via style transfer: a variational autoencoder for disentangling spatial images and musical content

In the stereo-to-multichannel upmixing problem for music, one of the mai...
research
06/25/2021

InteL-VAEs: Adding Inductive Biases to Variational Auto-Encoders via Intermediary Latents

We introduce a simple and effective method for learning VAEs with contro...
research
02/10/2022

Translation and Rotation Equivariant Normalizing Flow (TRENF) for Optimal Cosmological Analysis

Our universe is homogeneous and isotropic, and its perturbations obey tr...
research
06/29/2020

VAE-KRnet and its applications to variational Bayes

In this work, we have proposed a generative model for density estimation...
research
01/10/2021

Learning Rotation Invariant Features for Cryogenic Electron Microscopy Image Reconstruction

Cryo-Electron Microscopy (Cryo-EM) is a Nobel prize-winning technology f...
research
09/07/2020

Ordinal-Content VAE: Isolating Ordinal-Valued Content Factors in Deep Latent Variable Models

In deep representational learning, it is often desired to isolate a part...

Please sign up or login with your details

Forgot password? Click here to reset