Label Noise: Correcting a Correction

07/24/2023
by   William Toner, et al.
0

Training neural network classifiers on datasets with label noise poses a risk of overfitting them to the noisy labels. To address this issue, researchers have explored alternative loss functions that aim to be more robust. However, many of these alternatives are heuristic in nature and still vulnerable to overfitting or underfitting. In this work, we propose a more direct approach to tackling overfitting caused by label noise. We observe that the presence of label noise implies a lower bound on the noisy generalised risk. Building upon this observation, we propose imposing a lower bound on the empirical risk during training to mitigate overfitting. Our main contribution is providing theoretical results that yield explicit, easily computable bounds on the minimum achievable noisy risk for different loss functions. We empirically demonstrate that using these bounds significantly enhances robustness in various settings, with virtually no additional computational cost.

READ FULL TEXT

page 9

page 19

research
10/07/2021

Robustness and reliability when training with noisy labels

Labelling of data for supervised learning can be costly and time-consumi...
research
06/06/2023

Binary Classification with Instance and Label Dependent Label Noise

Learning with label dependent label noise has been extensively explored ...
research
02/16/2020

Learning Not to Learn in the Presence of Noisy Labels

Learning in the presence of label noise is a challenging yet important t...
research
05/03/2023

A Curriculum View of Robust Loss Functions

Robust loss functions are designed to combat the adverse impacts of labe...
research
05/23/2023

Mitigating Label Noise through Data Ambiguation

Label noise poses an important challenge in machine learning, especially...
research
04/06/2023

Logistic-Normal Likelihoods for Heteroscedastic Label Noise in Classification

A natural way of estimating heteroscedastic label noise in regression is...
research
10/06/2021

Foolish Crowds Support Benign Overfitting

We prove a lower bound on the excess risk of sparse interpolating proced...

Please sign up or login with your details

Forgot password? Click here to reset