Tackling Instance-Dependent Label Noise with Dynamic Distribution Calibration

10/11/2022
by   Manyi Zhang, et al.
0

Instance-dependent label noise is realistic but rather challenging, where the label-corruption process depends on instances directly. It causes a severe distribution shift between the distributions of training and test data, which impairs the generalization of trained models. Prior works put great effort into tackling the issue. Unfortunately, these works always highly rely on strong assumptions or remain heuristic without theoretical guarantees. In this paper, to address the distribution shift in learning with instance-dependent label noise, a dynamic distribution-calibration strategy is adopted. Specifically, we hypothesize that, before training data are corrupted by label noise, each class conforms to a multivariate Gaussian distribution at the feature level. Label noise produces outliers to shift the Gaussian distribution. During training, to calibrate the shifted distribution, we propose two methods based on the mean and covariance of multivariate Gaussian distribution respectively. The mean-based method works in a recursive dimension-reduction manner for robust mean estimation, which is theoretically guaranteed to train a high-quality model against label noise. The covariance-based method works in a distribution disturbance manner, which is experimentally verified to improve the model robustness. We demonstrate the utility and effectiveness of our methods on datasets with synthetic label noise and real-world unknown noise.

READ FULL TEXT
research
01/11/2020

Confidence Scores Make Instance-dependent Label-noise Learning Possible

Learning with noisy labels has drawn a lot of attention. In this area, m...
research
06/14/2020

Parts-dependent Label Noise: Towards Instance-dependent Label Noise

Learning with the instance-dependent label noise is challenging, because...
research
12/22/2020

A Second-Order Approach to Learning with Instance-Dependent Label Noise

The presence of label noise often misleads the training of deep neural n...
research
10/15/2022

Label distribution learning via label correlation grid

Label distribution learning can characterize the polysemy of an instance...
research
09/28/2022

Label Distribution Learning via Implicit Distribution Representation

In contrast to multi-label learning, label distribution learning charact...
research
12/20/2022

Improving the Robustness of Summarization Models by Detecting and Removing Input Noise

The evaluation of abstractive summarization models typically uses test d...
research
03/13/2021

Learning with Feature-Dependent Label Noise: A Progressive Approach

Label noise is frequently observed in real-world large-scale datasets. T...

Please sign up or login with your details

Forgot password? Click here to reset