Class Prototype-based Cleaner for Label Noise Learning

12/21/2022
by   Jingjia Huang, et al.
0

Semi-supervised learning based methods are current SOTA solutions to the noisy-label learning problem, which rely on learning an unsupervised label cleaner first to divide the training samples into a labeled set for clean data and an unlabeled set for noise data. Typically, the cleaner is obtained via fitting a mixture model to the distribution of per-sample training losses. However, the modeling procedure is class agnostic and assumes the loss distributions of clean and noise samples are the same across different classes. Unfortunately, in practice, such an assumption does not always hold due to the varying learning difficulty of different classes, thus leading to sub-optimal label noise partition criteria. In this work, we reveal this long-ignored problem and propose a simple yet effective solution, named Class Prototype-based label noise Cleaner (CPC). Unlike previous works treating all the classes equally, CPC fully considers loss distribution heterogeneity and applies class-aware modulation to partition the clean and noise data. CPC takes advantage of loss distribution modeling and intra-class consistency regularization in feature space simultaneously and thus can better distinguish clean and noise labels. We theoretically justify the effectiveness of our method by explaining it from the Expectation-Maximization (EM) framework. Extensive experiments are conducted on the noisy-label benchmarks CIFAR-10, CIFAR-100, Clothing1M and WebVision. The results show that CPC consistently brings about performance improvement across all benchmarks. Codes and pre-trained models will be released at <https://github.com/hjjpku/CPC.git>.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/13/2023

Manifold DivideMix: A Semi-Supervised Contrastive Learning Framework for Severe Label Noise

Deep neural networks have proven to be highly effective when large amoun...
research
08/21/2022

Combating Noisy-Labeled and Imbalanced Data by Two Stage Bi-Dimensional Sample Selection

Robust learning on noisy-labeled data has been an important task in real...
research
07/21/2022

ProMix: Combating Label Noise via Maximizing Clean Sample Utility

The ability to train deep neural networks under label noise is appealing...
research
01/04/2023

Towards the Identifiability in Noisy Label Learning: A Multinomial Mixture Approach

Learning from noisy labels plays an important role in the deep learning ...
research
12/31/2022

Unlearnable Clusters: Towards Label-agnostic Unlearnable Examples

There is a growing interest in developing unlearnable examples (UEs) aga...
research
08/29/2023

Prototype Fission: Closing Set for Robust Open-set Semi-supervised Learning

Semi-supervised Learning (SSL) has been proven vulnerable to out-of-dist...
research
05/29/2023

ReSup: Reliable Label Noise Suppression for Facial Expression Recognition

Because of the ambiguous and subjective property of the facial expressio...

Please sign up or login with your details

Forgot password? Click here to reset