Feature Lenses: Plug-and-play Neural Modules for Transformation-Invariant Visual Representations

04/12/2020
by   Shaohua Li, et al.
0

Convolutional Neural Networks (CNNs) are known to be brittle under various image transformations, including rotations, scalings, and changes of lighting conditions. We observe that the features of a transformed image are drastically different from the ones of the original image. To make CNNs more invariant to transformations, we propose "Feature Lenses", a set of ad-hoc modules that can be easily plugged into a trained model (referred to as the "host model"). Each individual lens reconstructs the original features given the features of a transformed image under a particular transformation. These lenses jointly counteract feature distortions caused by various transformations, thus making the host model more robust without retraining. By only updating lenses, the host model is freed from iterative updating when facing new transformations absent in the training data; as feature semantics are preserved, downstream applications, such as classifiers and detectors, automatically gain robustness without retraining. Lenses are trained in a self-supervised fashion with no annotations, by minimizing a novel "Top-K Activation Contrast Loss" between lens-transformed features and original features. Evaluated on ImageNet, MNIST-rot, and CIFAR-10, Feature Lenses show clear advantages over baseline methods.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/30/2020

Is Robustness To Transformations Driven by Invariant Neural Representations?

Deep Convolutional Neural Networks (DCNNs) have demonstrated impressive ...
research
02/08/2022

TransformNet: Self-supervised representation learning through predicting geometric transformations

Deep neural networks need a big amount of training data, while in the re...
research
10/03/2018

PADDIT: Probabilistic Augmentation of Data using Diffeomorphic Image Transformation

For proper generalization performance of convolutional neural networks (...
research
03/23/2019

AVT: Unsupervised Learning of Transformation Equivariant Representations by Autoencoding Variational Transformations

The learning of Transformation-Equivariant Representations (TERs), which...
research
06/19/2019

Learning Generalized Transformation Equivariant Representations via Autoencoding Transformations

Learning Transformation Equivariant Representations (TERs) seeks to capt...
research
11/25/2021

Quantised Transforming Auto-Encoders: Achieving Equivariance to Arbitrary Transformations in Deep Networks

In this work we investigate how to achieve equivariance to input transfo...
research
06/19/2020

Deep Transformation-Invariant Clustering

Recent advances in image clustering typically focus on learning better d...

Please sign up or login with your details

Forgot password? Click here to reset