Learning with Instance-Dependent Label Noise: A Sample Sieve Approach

10/05/2020
by   Hao Cheng, et al.
1

Human-annotated labels are often prone to noise, and the presence of such noise will degrade the performance of the resulting deep neural network (DNN) models. Much of the literature (with several recent exceptions) of learning with noisy labels focuses on the case when the label noise is independent from features. Practically, annotations errors tend to be instance-dependent and often depend on the difficulty levels of recognizing a certain task. Applying existing results from instance-independent settings would require a significant amount of estimation of noise rates. Therefore, learning with instance-dependent label noise remains a challenge. In this paper, we propose CORES^2 (COnfidence REgularized Sample Sieve), which progressively sieves out corrupted samples. The implementation of CORES^2 does not require specifying noise rates and yet we are able to provide theoretical guarantees of CORES^2 in filtering out the corrupted examples. This high-quality sample sieve allows us to treat clean examples and the corrupted ones separately in training a DNN solution, and such a separation is shown to be advantageous in the instance-dependent noise setting. We demonstrate the performance of CORES^2 on CIFAR10 and CIFAR100 datasets with synthetic instance-dependent label noise and Clothing1M with real-world human noise. As of independent interests, our sample sieve provides a generic machinery for anatomizing noisy datasets and provides a flexible interface for various robust training techniques to further improve the performance.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/22/2020

A Second-Order Approach to Learning with Instance-Dependent Label Noise

The presence of label noise often misleads the training of deep neural n...
research
11/29/2022

On Robust Learning from Noisy Labels: A Permutation Layer Approach

The existence of label noise imposes significant challenges (e.g., poor ...
research
03/13/2021

Learning with Feature-Dependent Label Noise: A Progressive Approach

Label noise is frequently observed in real-world large-scale datasets. T...
research
05/28/2021

Rethinking Noisy Label Models: Labeler-Dependent Noise with Adversarial Awareness

Most studies on learning from noisy labels rely on unrealistic models of...
research
07/16/2021

When does loss-based prioritization fail?

Not all examples are created equal, but standard deep neural network tra...
research
06/14/2019

Deep neural network for fringe pattern filtering and normalisation

We propose a new framework for processing Fringe Patterns (FP). Our nove...
research
03/25/2021

Approximating Instance-Dependent Noise via Instance-Confidence Embedding

Label noise in multiclass classification is a major obstacle to the depl...

Please sign up or login with your details

Forgot password? Click here to reset