An Explicit Local and Global Representation Disentanglement Framework with Applications in Deep Clustering and Unsupervised Object Detection

01/24/2020
by   Rujikorn Charakorn, et al.
5

There are several benefits from learning disentangled representations, including interpretability, compositionality and generalisation to new tasks. Disentanglement could be done by imposing an inductive bias based on prior knowledge. Different priors make different structural assumptions about the representations. For instance, priors with granularity can lead to representation to describe data at different scales. A notable example is in the visual domain where there are multiple scales of the variation in the data. Hence, learning representation at different scales will be useful for different tasks. In this work, we propose a framework, called SPLIT, which allows us to disentangle local and global information into two separate sets of latent variables within the variational autoencoder (VAE) framework. Our framework adds an extra generative assumption to the VAE by requiring a subset of the latent variables to generate an auxiliary set of observable data. This set of data, which contains only local information, can be obtained via a transformation of the original data that removes global information. Three different flavours of VAE with different generative assumptions were examined in our experiments. We show that the framework can be effectively used to disentangle local and global information within these models. Benefits of the framework are demonstrated through multiple downstream representation learning problems. The framework can unlock the potential of these VAE models in the tasks of style transfer, deep clustering and unsupervised object detection with a simple modification to existing VAE models. Finally, we review cognitive neuroscience literature regarding disentanglement in human visual perception. The code for our experiments can be found at https://github.com/51616/split-vae.

READ FULL TEXT

page 1

page 4

page 6

page 7

page 8

page 9

page 13

research
11/08/2016

Variational Lossy Autoencoder

Representation learning seeks to expose certain aspects of observed data...
research
03/29/2021

SetVAE: Learning Hierarchical Composition for Generative Modeling of Set-Structured Data

Generative modeling of set-structured data, such as point clouds, requir...
research
06/25/2021

InteL-VAEs: Adding Inductive Biases to Variational Auto-Encoders via Intermediary Latents

We introduce a simple and effective method for learning VAEs with contro...
research
05/27/2019

Wyner VAE: Joint and Conditional Generation with Succinct Common Representation Learning

A new variational autoencoder (VAE) model is proposed that learns a succ...
research
01/24/2022

A Bayesian Permutation training deep representation learning method for speech enhancement with variational autoencoder

Recently, variational autoencoder (VAE), a deep representation learning ...
research
06/17/2016

Early Visual Concept Learning with Unsupervised Deep Learning

Automated discovery of early visual concepts from raw image data is a ma...
research
07/27/2023

Online Clustered Codebook

Vector Quantisation (VQ) is experiencing a comeback in machine learning,...

Please sign up or login with your details

Forgot password? Click here to reset