A Semi-Supervised Two-Stage Approach to Learning from Noisy Labels

02/08/2018
by   Yifan Ding, et al.
0

The recent success of deep neural networks is powered in part by large-scale well-labeled training data. However, it is a daunting task to laboriously annotate an ImageNet-like dateset. On the contrary, it is fairly convenient, fast, and cheap to collect training images from the Web along with their noisy labels. This signifies the need of alternative approaches to training deep neural networks using such noisy labels. Existing methods tackling this problem either try to identify and correct the wrong labels or reweigh the data terms in the loss function according to the inferred noisy rates. Both strategies inevitably incur errors for some of the data points. In this paper, we contend that it is actually better to ignore the labels of some of the data points than to keep them if the labels are incorrect, especially when the noisy rate is high. After all, the wrong labels could mislead a neural network to a bad local optimum. We suggest a two-stage framework for the learning from noisy labels. In the first stage, we identify a small portion of images from the noisy training set of which the labels are correct with a high probability. The noisy labels of the other images are ignored. In the second stage, we train a deep neural network in a semi-supervised manner. This framework effectively takes advantage of the whole training set and yet only a portion of its labels that are most likely correct. Experiments on three datasets verify the effectiveness of our approach especially when the noisy rate is high.

READ FULL TEXT
research
10/07/2016

Temporal Ensembling for Semi-Supervised Learning

In this paper, we present a simple and efficient method for training dee...
research
12/06/2021

Two Wrongs Don't Make a Right: Combating Confirmation Bias in Learning with Label Noise

Noisy labels damage the performance of deep networks. For robust learnin...
research
03/01/2021

DST: Data Selection and joint Training for Learning with Noisy Labels

Training a deep neural network heavily relies on a large amount of train...
research
10/17/2021

Data Shapley Value for Handling Noisy Labels: An application in Screening COVID-19 Pneumonia from Chest CT Scans

A long-standing challenge of deep learning models involves how to handle...
research
08/03/2018

CurriculumNet: Weakly Supervised Learning from Large-Scale Web Images

We present a simple yet efficient approach capable of training deep neur...
research
11/22/2022

OR-Gate: A Noisy Label Filtering Method for Speaker Verification

The deep learning models used for speaker verification are heavily depen...
research
03/15/2022

Scalable Penalized Regression for Noise Detection in Learning with Noisy Labels

Noisy training set usually leads to the degradation of generalization an...

Please sign up or login with your details

Forgot password? Click here to reset