PropMix: Hard Sample Filtering and Proportional MixUp for Learning with Noisy Labels

10/22/2021
by   Filipe R. Cordeiro, et al.
10

The most competitive noisy label learning methods rely on an unsupervised classification of clean and noisy samples, where samples classified as noisy are re-labelled and "MixMatched" with the clean samples. These methods have two issues in large noise rate problems: 1) the noisy set is more likely to contain hard samples that are in-correctly re-labelled, and 2) the number of samples produced by MixMatch tends to be reduced because it is constrained by the small clean set size. In this paper, we introduce the learning algorithm PropMix to handle the issues above. PropMix filters out hard noisy samples, with the goal of increasing the likelihood of correctly re-labelling the easy noisy samples. Also, PropMix places clean and re-labelled easy noisy samples in a training set that is augmented with MixUp, removing the clean set size constraint and including a large proportion of correctly re-labelled easy noisy samples. We also include self-supervised pre-training to improve robustness to high noisy label scenarios. Our experiments show that PropMix has state-of-the-art (SOTA) results on CIFAR-10/-100(with symmetric, asymmetric and semantic label noise), Red Mini-ImageNet (from the Controlled Noisy Web Labels), Clothing1M and WebVision. In severe label noise bench-marks, our results are substantially better than other methods. The code is available athttps://github.com/filipe-research/PropMix.

READ FULL TEXT
research
03/06/2021

LongReMix: Robust Learning with High Confidence Samples in a Noisy Label Environment

Deep neural network models are robust to a limited amount of label noise...
research
03/20/2023

PASS: Peer-Agreement based Sample Selection for training with Noisy Labels

Noisy labels present a significant challenge in deep learning because mo...
research
06/20/2023

LNL+K: Learning with Noisy Labels and Noise Source Distribution Knowledge

Learning with noisy labels (LNL) is challenging as the model tends to me...
research
02/04/2022

Learning with Neighbor Consistency for Noisy Labels

Recent advances in deep learning have relied on large, labelled datasets...
research
08/26/2023

Late Stopping: Avoiding Confidently Learning from Mislabeled Examples

Sample selection is a prevalent method in learning with noisy labels, wh...
research
10/19/2020

GANs for learning from very high class conditional noisy labels

We use Generative Adversarial Networks (GANs) to design a class conditio...
research
11/20/2018

Limited Gradient Descent: Learning With Noisy Labels

Label noise may handicap the generalization of classifiers, and it is an...

Please sign up or login with your details

Forgot password? Click here to reset