Sparse Spatial Transformers for Few-Shot Learning

09/27/2021
by   Haoxing Chen, et al.
0

Learning from limited data is a challenging task since the scarcity of data leads to a poor generalization of the trained model. The classical global pooled representation is likely to lose useful local information. Recently, many few shot learning methods address this challenge by using deep descriptors and learning a pixel-level metric. However, using deep descriptors as feature representations may lose the contextual information of the image. And most of these methods deal with each class in the support set independently, which cannot sufficiently utilize discriminative information and task-specific embeddings. In this paper, we propose a novel Transformer based neural network architecture called Sparse Spatial Transformers (SSFormers), which can find task-relevant features and suppress task-irrelevant features. Specifically, we first divide each input image into several image patches of different sizes to obtain dense local features. These features retain contextual information while expressing local information. Then, a sparse spatial transformer layer is proposed to find spatial correspondence between the query image and the entire support set to select task-relevant image patches and suppress task-irrelevant image patches. Finally, we propose an image patch matching module to calculate the distance between dense local representations to determine which category the query image belongs to in the support set. Extensive experiments on popular few-shot learning benchmarks show that our method achieves the state-of-the-art performance. Our code is available at <https://github.com/chenhaoxing/SSFormers>.

READ FULL TEXT

page 6

page 7

research
04/26/2023

ESPT: A Self-Supervised Episodic Spatial Pretext Task for Improving Few-Shot Learning

Self-supervised learning (SSL) techniques have recently been integrated ...
research
10/22/2021

GCCN: Global Context Convolutional Network

In this paper, we propose Global Context Convolutional Network (GCCN) fo...
research
08/26/2022

Few-Shot Learning Meets Transformer: Unified Query-Support Transformers for Few-Shot Classification

Few-shot classification which aims to recognize unseen classes using ver...
research
05/24/2017

Dense Transformer Networks

The key idea of current deep learning methods for dense prediction is to...
research
07/15/2020

Augmented Bi-path Network for Few-shot Learning

Few-shot Learning (FSL) which aims to learn from few labeled training da...
research
10/30/2022

Alleviating the Sample Selection Bias in Few-shot Learning by Removing Projection to the Centroid

Few-shot learning (FSL) targets at generalization of vision models towar...
research
11/01/2021

Few-shot learning with improved local representations via bias rectify module

Recent approaches based on metric learning have achieved great progress ...

Please sign up or login with your details

Forgot password? Click here to reset