Align, Distill, and Augment Everything All at Once for Imbalanced Semi-Supervised Learning

06/07/2023
by   Emanuel Sanchez Aimar, et al.
0

Addressing the class imbalance in long-tailed semi-supervised learning (SSL) poses a few significant challenges stemming from differences between the marginal distributions of unlabeled data and the labeled data, as the former is often unknown and potentially distinct from the latter. The first challenge is to avoid biasing the pseudo-labels towards an incorrect distribution, such as that of the labeled data or a balanced distribution, during training. However, we still wish to ensure a balanced unlabeled distribution during inference, which is the second challenge. To address both of these challenges, we propose a three-faceted solution: a flexible distribution alignment that progressively aligns the classifier from a dynamically estimated unlabeled prior towards a balanced distribution, a soft consistency regularization that exploits underconfident pseudo-labels discarded by threshold-based methods, and a schema for expanding the unlabeled set with input data from the labeled partition. This last facet comes in as a response to the commonly-overlooked fact that disjoint partitions of labeled and unlabeled data prevent the benefits of strong data augmentation on the labeled set. Our overall framework requires no additional training cycles, so it will align, distill, and augment everything all at once (ADALLO). Our extensive evaluations of ADALLO on imbalanced SSL benchmark datasets, including CIFAR10-LT, CIFAR100-LT, and STL10-LT with varying degrees of class imbalance, amount of labeled data, and distribution mismatch, demonstrate significant improvements in the performance of imbalanced SSL under large distribution mismatch, as well as competitiveness with state-of-the-art methods when the labeled and unlabeled data follow the same marginal distribution. Our code will be released upon paper acceptance.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/17/2020

Class-Imbalanced Semi-Supervised Learning

Semi-Supervised Learning (SSL) has achieved great success in overcoming ...
research
10/20/2021

ABC: Auxiliary Balanced Classifier for Class-imbalanced Semi-supervised Learning

Existing semi-supervised learning (SSL) algorithms typically assume clas...
research
05/22/2023

BMB: Balanced Memory Bank for Imbalanced Semi-supervised Learning

Exploring a substantial amount of unlabeled data, semi-supervised learni...
research
03/15/2020

Beyond without Forgetting: Multi-Task Learning for Classification with Disjoint Datasets

Multi-task Learning (MTL) for classification with disjoint datasets aims...
research
05/07/2022

Calibrating Label Distribution for Class-Imbalanced Barely-Supervised Knee Segmentation

Segmentation of 3D knee MR images is important for the assessment of ost...
research
04/07/2012

Density-sensitive semisupervised inference

Semisupervised methods are techniques for using labeled data (X_1,Y_1),....
research
08/07/2019

Unsupervised Feature Learning in Remote Sensing

The need for labeled data is among the most common and well-known practi...

Please sign up or login with your details

Forgot password? Click here to reset