Noisy Labels Can Induce Good Representations

12/23/2020
by   Jingling Li, et al.
11

The current success of deep learning depends on large-scale labeled datasets. In practice, high-quality annotations are expensive to collect, but noisy annotations are more affordable. Previous works report mixed empirical results when training with noisy labels: neural networks can easily memorize random labels, but they can also generalize from noisy labels. To explain this puzzle, we study how architecture affects learning with noisy labels. We observe that if an architecture "suits" the task, training with noisy labels can induce useful hidden representations, even when the model generalizes poorly; i.e., the last few layers of the model are more negatively affected by noisy labels. This finding leads to a simple method to improve models trained on noisy labels: replacing the final dense layers with a linear model, whose weights are learned from a small set of clean data. We empirically validate our findings across three architectures (Convolutional Neural Networks, Graph Neural Networks, and Multi-Layer Perceptrons) and two domains (graph algorithmic tasks and image classification). Furthermore, we achieve state-of-the-art results on image classification benchmarks by combining our method with existing approaches on noisy label training.

READ FULL TEXT

page 2

page 9

research
03/06/2021

Noisy Label Learning for Large-scale Medical Image Classification

The classification accuracy of deep learning models depends not only on ...
research
08/31/2019

Learning with Noisy Labels for Sentence-level Sentiment Classification

Deep neural networks (DNNs) can fit (or even over-fit) the training data...
research
05/12/2023

Expertise-based Weighting for Regression Models with Noisy Labels

Regression methods assume that accurate labels are available for trainin...
research
06/30/2021

Understanding and Improving Early Stopping for Learning with Noisy Labels

The memorization effect of deep neural network (DNN) plays a pivotal rol...
research
06/30/2020

Early-Learning Regularization Prevents Memorization of Noisy Labels

We propose a novel framework to perform classification via deep learning...
research
03/24/2020

Robust and On-the-fly Dataset Denoising for Image Classification

Memorization in over-parameterized neural networks could severely hurt g...
research
08/15/2019

Improved Mix-up with KL-Entropy for Learning From Noisy Labels

Despite the deep neural networks (DNN) has achieved excellent performanc...

Please sign up or login with your details

Forgot password? Click here to reset