Steerable Equivariant Representation Learning

02/22/2023
by   Sangnie Bhardwaj, et al.
0

Pre-trained deep image representations are useful for post-training tasks such as classification through transfer learning, image retrieval, and object detection. Data augmentations are a crucial aspect of pre-training robust representations in both supervised and self-supervised settings. Data augmentations explicitly or implicitly promote invariance in the embedding space to the input image transformations. This invariance reduces generalization to those downstream tasks which rely on sensitivity to these particular data augmentations. In this paper, we propose a method of learning representations that are instead equivariant to data augmentations. We achieve this equivariance through the use of steerable representations. Our representations can be manipulated directly in embedding space via learned linear maps. We demonstrate that our resulting steerable and equivariant representations lead to better performance on transfer learning and robustness: e.g. we improve linear probe top-1 accuracy by between 1 and ImageNet-C accuracy by upto 3.4 our representations provides significant speedup (nearly 50x) for test-time augmentations; by applying a large number of augmentations for out-of-distribution detection, we significantly improve OOD AUC on the ImageNet-C dataset over an invariant representation.

READ FULL TEXT

page 2

page 7

page 8

page 9

page 23

page 24

page 25

page 26

research
07/27/2022

On the robustness of self-supervised representations for multi-view object classification

It is known that representations from self-supervised pre-training can p...
research
06/28/2023

DUET: 2D Structured and Approximately Equivariant Representations

Multiview Self-Supervised Learning (MSSL) is based on learning invarianc...
research
10/12/2022

GULP: a prediction-based metric between representations

Comparing the representations learned by different neural networks has r...
research
10/30/2022

A simple, efficient and scalable contrastive masked autoencoder for learning visual representations

We introduce CAN, a simple, efficient and scalable method for self-super...
research
02/22/2021

On Interaction Between Augmentations and Corruptions in Natural Corruption Robustness

Invariance to a broad array of image corruptions, such as warping, noise...
research
09/29/2022

Towards General-Purpose Representation Learning of Polygonal Geometries

Neural network representation learning for spatial data is a common need...
research
06/24/2023

Structuring Representation Geometry with Rotationally Equivariant Contrastive Learning

Self-supervised learning converts raw perceptual data such as images to ...

Please sign up or login with your details

Forgot password? Click here to reset