On Emergence of Clean-Priority Learning in Early Stopped Neural Networks

06/05/2023
by   Chaoyue Liu, et al.
0

When random label noise is added to a training dataset, the prediction error of a neural network on a label-noise-free test dataset initially improves during early training but eventually deteriorates, following a U-shaped dependence on training time. This behaviour is believed to be a result of neural networks learning the pattern of clean data first and fitting the noise later in the training, a phenomenon that we refer to as clean-priority learning. In this study, we aim to explore the learning dynamics underlying this phenomenon. We theoretically demonstrate that, in the early stage of training, the update direction of gradient descent is determined by the clean subset of training data, leaving the noisy subset has minimal to no impact, resulting in a prioritization of clean learning. Moreover, we show both theoretically and experimentally, as the clean-priority learning goes on, the dominance of the gradients of clean samples over those of noisy samples diminishes, and finally results in a termination of the clean-priority learning and fitting of the noisy samples.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/20/2019

Leveraging inductive bias of neural networks for learning without explicit human annotations

Classification problems today are typically solved by first collecting e...
research
03/02/2023

Over-training with Mixup May Hurt Generalization

Mixup, which creates synthetic training instances by linearly interpolat...
research
12/24/2020

Identifying Training Stop Point with Noisy Labeled Data

Training deep neural networks (DNNs) with noisy labels is a challenging ...
research
07/21/2022

ProMix: Combating Label Noise via Maximizing Clean Sample Utility

The ability to train deep neural networks under label noise is appealing...
research
05/27/2021

Using Early-Learning Regularization to Classify Real-World Noisy Data

The memorization problem is well-known in the field of computer vision. ...
research
11/20/2018

Limited Gradient Descent: Learning With Noisy Labels

Label noise may handicap the generalization of classifiers, and it is an...
research
05/14/2019

Task-Driven Data Verification via Gradient Descent

We introduce a novel algorithm for the detection of possible sample corr...

Please sign up or login with your details

Forgot password? Click here to reset