Robustness and reliability when training with noisy labels

10/07/2021
by   Amanda Olmin, et al.
0

Labelling of data for supervised learning can be costly and time-consuming and the risk of incorporating label noise in large data sets is imminent. If training a flexible discriminative model using a strictly proper loss, such noise will inevitably shift the solution towards the conditional distribution over noisy labels. Nevertheless, while deep neural networks have proved capable of fitting random labels, regularisation and the use of robust loss functions empirically mitigate the effects of label noise. However, such observations concern robustness in accuracy, which is insufficient if reliable uncertainty quantification is critical. We demonstrate this by analysing the properties of the conditional distribution over noisy labels for an input-dependent noise model. In addition, we evaluate the set of robust loss functions characterised by an overlap in asymptotic risk minimisers under the clean and noisy data distributions. We find that strictly proper and robust loss functions both offer asymptotic robustness in accuracy, but neither guarantee that the resulting model is calibrated. Moreover, overfitting is an issue in practice. With these results, we aim to explain inherent robustness of algorithms to label noise and to give guidance in the development of new noise-robust algorithms.

READ FULL TEXT
research
12/27/2017

Robust Loss Functions under Label Noise for Deep Neural Networks

In many applications of classifier learning, training data suffers from ...
research
07/24/2023

Label Noise: Correcting a Correction

Training neural network classifiers on datasets with label noise poses a...
research
06/21/2018

Robust and Efficient Boosting Method using the Conditional Risk

Well-known for its simplicity and effectiveness in classification, AdaBo...
research
11/07/2020

When Optimizing f-divergence is Robust with Label Noise

We show when maximizing a properly defined f-divergence measure with res...
research
08/03/2022

Noise tolerance of learning to rank under class-conditional label noise

Often, the data used to train ranking models is subject to label noise. ...
research
04/06/2023

Logistic-Normal Likelihoods for Heteroscedastic Label Noise in Classification

A natural way of estimating heteroscedastic label noise in regression is...
research
04/03/2021

Exponentiated Gradient Reweighting for Robust Training Under Label Noise and Beyond

Many learning tasks in machine learning can be viewed as taking a gradie...

Please sign up or login with your details

Forgot password? Click here to reset