Exploring Feature Self-relation for Self-supervised Transformer

06/10/2022
by   Zhong-Yu Li, et al.
0

Learning representations with self-supervision for convolutional networks (CNN) has proven effective for vision tasks. As an alternative for CNN, vision transformers (ViTs) emerge strong representation ability with the pixel-level self-attention and channel-level feed-forward networks. Recent works reveal that self-supervised learning helps unleash the great potential of ViTs. Still, most works follow self-supervised strategy designed for CNNs, e.g., instance-level discrimination of samples, but they ignore the unique properties of ViTs. We observe that modeling relations among pixels and channels distinguishes ViTs from other networks. To enforce this property, we explore the feature self-relations for training self-supervised ViTs. Specifically, instead of conducting self-supervised learning solely on feature embeddings from multiple views, we utilize the feature self-relations, i.e., pixel/channel-level self-relations, for self-supervised learning. Self-relation based learning further enhance the relation modeling ability of ViTs, resulting in strong representations that stably improve performance on multiple downstream tasks. Our source code will be made publicly available.

READ FULL TEXT

page 2

page 5

page 12

research
05/10/2021

Self-Supervised Learning with Swin Transformers

We are witnessing a modeling shift from CNN to Transformers in computer ...
research
11/02/2022

Beyond Instance Discrimination: Relation-aware Contrastive Self-supervised Learning

Contrastive self-supervised learning (CSL) based on instance discriminat...
research
10/26/2021

Understanding the Role of Self-Supervised Learning in Out-of-Distribution Detection Task

Self-supervised learning (SSL) has achieved great success in a variety o...
research
04/06/2022

LEAD: Self-Supervised Landmark Estimation by Aligning Distributions of Feature Similarity

In this work, we introduce LEAD, an approach to discover landmarks from ...
research
07/18/2023

MOCA: Self-supervised Representation Learning by Predicting Masked Online Codebook Assignments

Self-supervised learning can be used for mitigating the greedy needs of ...
research
04/19/2022

Rumor Detection with Self-supervised Learning on Texts and Social Graph

Rumor detection has become an emerging and active research field in rece...
research
05/17/2023

Self-Supervised Learning for Physiologically-Based Pharmacokinetic Modeling in Dynamic PET

Dynamic positron emission tomography imaging (dPET) provides temporally ...

Please sign up or login with your details

Forgot password? Click here to reset