Towards Robust Learning with Different Label Noise Distributions

12/18/2019
by   Diego Ortego, et al.
20

Noisy labels are an unavoidable consequence of automatic image labeling processes to reduce human supervision. Training in these conditions leads Convolutional Neural Networks to memorize label noise and degrade performance. Noisy labels are therefore dispensable, while image content can be exploited in a semi-supervised learning (SSL) setup. Handling label noise then becomes a label noise detection task. Noisy/clean samples are usually identified using the small loss trick, which is based on the observation that clean samples represent easier patterns and, therefore, exhibit a lower loss. However, we show that different noise distributions make the application of this trick less straightforward. We propose to continuously relabel all images to reveal a loss that facilitates the use of the small loss trick with different noise distributions. SSL is then applied twice, once to improve the clean-noisy detection and again for training the final model. We design an experimental setup for better understanding the consequences of differing label noise distributions and find that non-uniform out-of-distribution noise better resembles real-world noise. We show that SSL outperforms other alternatives when using oracles and demonstrate substantial improvements across five datasets of our label noise Distribution Robust Pseudo-Labeling (DRPL). We further study the effects of label noise memorization via linear probes and find that in most cases intermediate features are not affected by label noise corruption. Code and details to reproduce our framework will be made available.

READ FULL TEXT

page 7

page 8

page 12

page 15

research
05/28/2023

BadLabel: A Robust Perspective on Evaluating and Enhancing Label-noise Learning

Label-noise learning (LNL) aims to increase the model's generalization g...
research
09/30/2020

Improving Generalization of Deep Fault Detection Models in the Presence of Mislabeled Data

Mislabeled samples are ubiquitous in real-world datasets as rule-based o...
research
10/26/2021

Addressing out-of-distribution label noise in webly-labelled data

A recurring focus of the deep learning community is towards reducing the...
research
09/04/2023

Robust Online Classification: From Estimation to Denoising

We study online classification in the presence of noisy labels. The nois...
research
06/20/2023

LNL+K: Learning with Noisy Labels and Noise Source Distribution Knowledge

Learning with noisy labels (LNL) is challenging as the model tends to me...
research
11/07/2020

When Optimizing f-divergence is Robust with Label Noise

We show when maximizing a properly defined f-divergence measure with res...
research
10/12/2020

Webly Supervised Image Classification with Metadata: Automatic Noisy Label Correction via Visual-Semantic Graph

Webly supervised learning becomes attractive recently for its efficiency...

Please sign up or login with your details

Forgot password? Click here to reset