Self Supervision to Distillation for Long-Tailed Visual Recognition

09/09/2021
by   Tianhao Li, et al.
0

Deep learning has achieved remarkable progress for visual recognition on large-scale balanced datasets but still performs poorly on real-world long-tailed data. Previous methods often adopt class re-balanced training strategies to effectively alleviate the imbalance issue, but might be a risk of over-fitting tail classes. The recent decoupling method overcomes over-fitting issues by using a multi-stage training scheme, yet, it is still incapable of capturing tail class information in the feature learning stage. In this paper, we show that soft label can serve as a powerful solution to incorporate label correlation into a multi-stage training scheme for long-tailed recognition. The intrinsic relation between classes embodied by soft labels turns out to be helpful for long-tailed recognition by transferring knowledge from head to tail classes. Specifically, we propose a conceptually simple yet particularly effective multi-stage training scheme, termed as Self Supervised to Distillation (SSD). This scheme is composed of two parts. First, we introduce a self-distillation framework for long-tailed recognition, which can mine the label relation automatically. Second, we present a new distillation label generation module guided by self-supervision. The distilled labels integrate information from both label and data domains that can model long-tailed distribution effectively. We conduct extensive experiments and our method achieves the state-of-the-art results on three long-tailed recognition benchmarks: ImageNet-LT, CIFAR100-LT and iNaturalist 2018. Our SSD outperforms the strong LWS baseline by from 2.7% to 4.5% on various datasets. The code is available at https://github.com/MCG-NJU/SSD-LT.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/12/2021

Class-Balanced Distillation for Long-Tailed Visual Recognition

Real-world imagery is often characterized by a significant imbalance of ...
research
05/15/2023

Global and Local Mixture Consistency Cumulative Learning for Long-tailed Visual Recognitions

In this paper, our goal is to design a simple learning paradigm for long...
research
07/19/2022

Invariant Feature Learning for Generalized Long-Tailed Classification

Existing long-tailed classification (LT) methods only focus on tackling ...
research
01/26/2021

ResLT: Residual Learning for Long-tailed Recognition

Deep learning algorithms face great challenges with long-tailed data dis...
research
11/26/2021

VL-LTR: Learning Class-wise Visual-Linguistic Representation for Long-Tailed Visual Recognition

Deep learning-based models encounter challenges when processing long-tai...
research
11/29/2021

A Simple Long-Tailed Recognition Baseline via Vision-Language Model

The visual world naturally exhibits a long-tailed distribution of open c...
research
02/06/2023

1st Place Solution for PSG competition with ECCV'22 SenseHuman Workshop

Panoptic Scene Graph (PSG) generation aims to generate scene graph repre...

Please sign up or login with your details

Forgot password? Click here to reset