Interpretable Transformations with Encoder-Decoder Networks

10/19/2017
by   Daniel E. Worrall, et al.
0

Deep feature spaces have the capacity to encode complex transformations of their input data. However, understanding the relative feature-space relationship between two transformed encoded images is difficult. For instance, what is the relative feature space relationship between two rotated images? What is decoded when we interpolate in feature space? Ideally, we want to disentangle confounding factors, such as pose, appearance, and illumination, from object identity. Disentangling these is difficult because they interact in very nonlinear ways. We propose a simple method to construct a deep feature space, with explicitly disentangled representations of several known transformations. A person or algorithm can then manipulate the disentangled representation, for example, to re-render an image with explicit control over parameterized degrees of freedom. The feature space is constructed using a transforming encoder-decoder network with a custom feature transform layer, acting on the hidden representations. We demonstrate the advantages of explicit disentangling on a variety of datasets and transformations, and as an aid for traditional tasks, such as classification.

READ FULL TEXT

page 7

page 8

research
04/24/2022

Learning Symmetric Embeddings for Equivariant World Models

Incorporating symmetries can lead to highly data-efficient and generaliz...
research
01/13/2018

Feature Space Transfer for Data Augmentation

The problem of data augmentation in feature space is considered. A new a...
research
12/13/2019

Laguerre-Gauss Preprocessing: Line Profiles as Image Features for Aerial Images Classification

An image preprocessing methodology based on Fourier analysis together wi...
research
09/21/2020

Learning Representation for Mixed Data Types with a Nonlinear Deep Encoder-Decoder Framework

Representation of data on mixed variables, numerical and categorical typ...
research
05/02/2023

Learning Disentangled Semantic Spaces of Explanations via Invertible Neural Networks

Disentangling sentence representations over continuous spaces can be a c...
research
10/15/2020

On Convergence of Nearest Neighbor Classifiers over Feature Transformations

The k-Nearest Neighbors (kNN) classifier is a fundamental non-parametric...
research
12/25/2019

Learning Controllable Disentangled Representations with Decorrelation Regularization

A crucial problem in learning disentangled image representations is cont...

Please sign up or login with your details

Forgot password? Click here to reset