Quantised Transforming Auto-Encoders: Achieving Equivariance to Arbitrary Transformations in Deep Networks

11/25/2021
by   Jianbo Jiao, et al.
0

In this work we investigate how to achieve equivariance to input transformations in deep networks, purely from data, without being given a model of those transformations. Convolutional Neural Networks (CNNs), for example, are equivariant to image translation, a transformation that can be easily modelled (by shifting the pixels vertically or horizontally). Other transformations, such as out-of-plane rotations, do not admit a simple analytic model. We propose an auto-encoder architecture whose embedding obeys an arbitrary set of equivariance relations simultaneously, such as translation, rotation, colour changes, and many others. This means that it can take an input image, and produce versions transformed by a given amount that were not observed before (e.g. a different point of view of the same object, or a colour variation). Despite extending to many (even non-geometric) transformations, our model reduces exactly to a CNN in the special case of translation-equivariance. Equivariances are important for the interpretability and robustness of deep networks, and we demonstrate results of successful re-rendering of transformed versions of input images on several synthetic and real datasets, as well as results on object pose estimation.

READ FULL TEXT

page 1

page 7

page 8

page 9

research
11/24/2017

Geometric robustness of deep networks: analysis and improvement

Deep convolutional neural networks have been shown to be vulnerable to a...
research
11/23/2015

What Happened to My Dog in That Network: Unraveling Top-down Generators in Convolutional Neural Networks

Top-down information plays a central role in human perception, but plays...
research
11/09/2022

ParGAN: Learning Real Parametrizable Transformations

Current methods for image-to-image translation produce compelling result...
research
06/13/2023

Effects of Data Enrichment with Image Transformations on the Performance of Deep Networks

Images cannot always be expected to come in a certain standard format an...
research
07/29/2014

How Auto-Encoders Could Provide Credit Assignment in Deep Networks via Target Propagation

We propose to exploit reconstruction as a layer-local training signal f...
research
12/04/2020

Learning Equivariant Representations

State-of-the-art deep learning systems often require large amounts of da...
research
04/12/2020

Feature Lenses: Plug-and-play Neural Modules for Transformation-Invariant Visual Representations

Convolutional Neural Networks (CNNs) are known to be brittle under vario...

Please sign up or login with your details

Forgot password? Click here to reset