CrossTransformers: spatially-aware few-shot transfer

07/22/2020
by   Carl Doersch, et al.
1

Given new tasks with very little data–such as new classes in a classification problem or a domain shift in the input–performance of modern vision systems degrades remarkably quickly. In this work, we illustrate how the neural network representations which underpin modern vision systems are subject to supervision collapse, whereby they lose any information that is not necessary for performing the training task, including information that may be necessary for transfer to new tasks or domains. We then propose two methods to mitigate this problem. First, we employ self-supervised learning to encourage general-purpose features that transfer better. Second, we propose a novel Transformer based neural network architecture called CrossTransformers, which can take a small number of labeled images and an unlabeled query, find coarse spatial correspondence between the query and the labeled images, and then infer class membership by computing distances between spatially-corresponding features. The result is a classifier that is more robust to task and domain shift, which we demonstrate via state-of-the-art performance on Meta-Dataset, a recent dataset for evaluating transfer from ImageNet to many other vision datasets.

READ FULL TEXT

page 2

page 6

page 17

page 18

page 19

page 20

page 21

research
10/08/2019

When Does Self-supervision Improve Few-shot Learning?

We present a technique to improve the generalization of deep representat...
research
06/19/2020

Self-Supervised Prototypical Transfer Learning for Few-Shot Classification

Most approaches in few-shot learning rely on costly annotated data relat...
research
06/16/2022

Adapting Self-Supervised Vision Transformers by Probing Attention-Conditioned Masking Consistency

Visual domain adaptation (DA) seeks to transfer trained models to unseen...
research
02/06/2018

Learning Image Representations by Completing Damaged Jigsaw Puzzles

In this paper, we explore methods of complicating self-supervised tasks ...
research
09/28/2020

Sense and Learn: Self-Supervision for Omnipresent Sensors

Learning general-purpose representations from multisensor data produced ...
research
02/07/2022

Reasoning for Complex Data through Ensemble-based Self-Supervised Learning

Self-supervised learning deals with problems that have little or no avai...
research
03/14/2023

OVRL-V2: A simple state-of-art baseline for ImageNav and ObjectNav

We present a single neural network architecture composed of task-agnosti...

Please sign up or login with your details

Forgot password? Click here to reset