Is your noise correction noisy? PLS: Robustness to label noise with two stage detection

10/10/2022
by   Paul Albert, et al.
20

Designing robust algorithms capable of training accurate neural networks on uncurated datasets from the web has been the subject of much research as it reduces the need for time consuming human labor. The focus of many previous research contributions has been on the detection of different types of label noise; however, this paper proposes to improve the correction accuracy of noisy samples once they have been detected. In many state-of-the-art contributions, a two phase approach is adopted where the noisy samples are detected before guessing a corrected pseudo-label in a semi-supervised fashion. The guessed pseudo-labels are then used in the supervised objective without ensuring that the label guess is likely to be correct. This can lead to confirmation bias, which reduces the noise robustness. Here we propose the pseudo-loss, a simple metric that we find to be strongly correlated with pseudo-label correctness on noisy samples. Using the pseudo-loss, we dynamically down weight under-confident pseudo-labels throughout training to avoid confirmation bias and improve the network accuracy. We additionally propose to use a confidence guided contrastive objective that learns robust representation on an interpolated objective between class bound (supervised) for confidently corrected samples and unsupervised representation for under-confident label corrections. Experiments demonstrate the state-of-the-art performance of our Pseudo-Loss Selection (PLS) algorithm on a variety of benchmark datasets including curated data synthetically corrupted with in-distribution and out-of-distribution noise, and two real world web noise datasets. Our experiments are fully reproducible [github coming soon]

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/26/2022

PARS: Pseudo-Label Aware Robust Sample Selection for Learning with Noisy Labels

Acquiring accurate labels on large-scale datasets is both time consuming...
research
12/06/2021

Two Wrongs Don't Make a Right: Combating Confirmation Bias in Learning with Label Noise

Noisy labels damage the performance of deep networks. For robust learnin...
research
10/26/2021

Addressing out-of-distribution label noise in webly-labelled data

A recurring focus of the deep learning community is towards reducing the...
research
07/07/2022

Semi-supervised Object Detection via Virtual Category Learning

Due to the costliness of labelled data in real-world applications, semi-...
research
02/09/2022

Learning to Bootstrap for Combating Label Noise

Deep neural networks are powerful tools for representation learning, but...
research
07/04/2022

Embedding contrastive unsupervised features to cluster in- and out-of-distribution noise in corrupted image datasets

Using search engines for web image retrieval is a tempting alternative t...
research
08/17/2023

URL: Combating Label Noise for Lung Nodule Malignancy Grading

Due to the complexity of annotation and inter-annotator variability, mos...

Please sign up or login with your details

Forgot password? Click here to reset