Adapting Self-Supervised Vision Transformers by Probing Attention-Conditioned Masking Consistency

06/16/2022
by   Viraj Prabhu, et al.
0

Visual domain adaptation (DA) seeks to transfer trained models to unseen, unlabeled domains across distribution shift, but approaches typically focus on adapting convolutional neural network architectures initialized with supervised ImageNet representations. In this work, we shift focus to adapting modern architectures for object recognition – the increasingly popular Vision Transformer (ViT) – and modern pretraining based on self-supervised learning (SSL). Inspired by the design of recent SSL approaches based on learning from partial image inputs generated via masking or cropping – either by learning to predict the missing pixels, or learning representational invariances to such augmentations – we propose PACMAC, a simple two-stage adaptation algorithm for self-supervised ViTs. PACMAC first performs in-domain SSL on pooled source and target data to learn task-discriminative features, and then probes the model's predictive consistency across a set of partial target inputs generated via a novel attention-conditioned masking strategy, to identify reliable candidates for self-training. Our simple approach leads to consistent performance gains over competing methods that use ViTs and self-supervised initializations on standard object recognition benchmarks. Code available at https://github.com/virajprabhu/PACMAC

READ FULL TEXT

page 4

page 9

page 14

research
07/25/2019

Self-supervised Domain Adaptation for Computer Vision Tasks

Recent progress of self-supervised visual representation learning has ac...
research
03/28/2020

Exploit Clues from Views: Self-Supervised and Regularized Learning for Multiview Object Recognition

Multiview recognition has been well studied in the literature and achiev...
research
10/15/2020

Self-Supervised Domain Adaptation with Consistency Training

We consider the problem of unsupervised domain adaptation for image clas...
research
09/21/2014

Domain Adaptive Neural Networks for Object Recognition

We propose a simple neural network model to deal with the domain adaptat...
research
07/22/2020

CrossTransformers: spatially-aware few-shot transfer

Given new tasks with very little data–such as new classes in a classific...
research
08/09/2017

Weakly- and Self-Supervised Learning for Content-Aware Deep Image Retargeting

This paper proposes a weakly- and self-supervised deep convolutional neu...
research
08/09/2023

GeoAdapt: Self-Supervised Test-Time Adaption in LiDAR Place Recognition Using Geometric Priors

LiDAR place recognition approaches based on deep learning suffer a signi...

Please sign up or login with your details

Forgot password? Click here to reset