Peer Loss Functions: Learning from Noisy Labels without Knowing Noise Rates

10/08/2019
by   Yang Liu, et al.
0

Learning with noisy labels is a common problem in supervised learning. Existing approaches require practitioners to specify noise rates, i.e., a set of parameters controlling the severity of label noises in the problem. In this work, we introduce a technique to learn from noisy labels that does not require a priori specification of the noise rates. In particular, we introduce a new family of loss functions that we name as peer loss functions. Our approach then uses a standard empirical risk minimization (ERM) framework with peer loss functions. Peer loss functions associate each training sample with a certain form of "peer" samples, which evaluate a classifier' predictions jointly. We show that, under mild conditions, performing ERM with peer loss functions on the noisy dataset leads to the optimal or a near optimal classifier as if performing ERM over the clean training data, which we do not have access to. To our best knowledge, this is the first result on "learning with noisy labels without knowing noise rates" with theoretical guarantees. We pair our results with an extensive set of experiments, where we compare with state-of-the-art techniques of learning with noisy labels. Our results show that peer loss functions based method consistently outperforms the baseline benchmarks. Peer loss provides a way to simplify model development when facing potentially noisy training labels, and can be promoted as a robust candidate loss function in such situations.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/27/2017

Robust Loss Functions under Label Noise for Deep Neural Networks

In many applications of classifier learning, training data suffers from ...
research
06/06/2021

Asymmetric Loss Functions for Learning with Noisy Labels

Robust loss functions are essential for training deep neural networks wi...
research
07/13/2019

Crisis: Probabilistically Self Organizing Total Order in Unstructured P2P Networks

A framework for asynchronous, signature free, fully local and probabilis...
research
03/08/2021

Program Synthesis Over Noisy Data with Guarantees

We explore and formalize the task of synthesizing programs over noisy da...
research
03/25/2021

Exploiting Class Similarity for Machine Learning with Confidence Labels and Projective Loss Functions

Class labels used for machine learning are relatable to each other, with...
research
04/13/2023

Bayes classifier cannot be learned from noisy responses with unknown noise rates

Training a classifier with noisy labels typically requires the learner t...
research
01/15/2020

Generalized Bayesian Quantification Learning

Quantification Learning is the task of prevalence estimation for a test ...

Please sign up or login with your details

Forgot password? Click here to reset