Instance-based Label Smoothing For Better Calibrated Classification Networks

10/11/2021
by   Mohamed Maher, et al.
0

Label smoothing is widely used in deep neural networks for multi-class classification. While it enhances model generalization and reduces overconfidence by aiming to lower the probability for the predicted class, it distorts the predicted probabilities of other classes resulting in poor class-wise calibration. Another method for enhancing model generalization is self-distillation where the predictions of a teacher network trained with one-hot labels are used as the target for training a student network. We take inspiration from both label smoothing and self-distillation and propose two novel instance-based label smoothing approaches, where a teacher network trained with hard one-hot labels is used to determine the amount of per class smoothness applied to each instance. The assigned smoothing factor is non-uniformly distributed along with the classes according to their similarity with the actual class. Our methods show better generalization and calibration over standard label smoothing on various deep neural architectures and image classification datasets.

READ FULL TEXT
research
06/06/2019

When Does Label Smoothing Help?

The generalization and learning speed of a multi-class neural network ca...
research
06/09/2020

Self-Distillation as Instance-Specific Label Smoothing

It has been recently demonstrated that multi-generational self-distillat...
research
06/24/2020

Class-Similarity Based Label Smoothing for Generalized Confidence Calibration

Since modern neural networks are known to be overconfident, several tech...
research
10/22/2022

Adaptive Label Smoothing with Self-Knowledge in Natural Language Generation

Overconfidence has been shown to impair generalization and calibration o...
research
09/11/2021

Class-Distribution-Aware Calibration for Long-Tailed Visual Recognition

Despite impressive accuracy, deep neural networks are often miscalibrate...
research
01/29/2023

Confidence-Aware Calibration and Scoring Functions for Curriculum Learning

Despite the great success of state-of-the-art deep neural networks, seve...
research
05/27/2019

On Mixup Training: Improved Calibration and Predictive Uncertainty for Deep Neural Networks

Mixup zhang2017mixup is a recently proposed method for training deep neu...

Please sign up or login with your details

Forgot password? Click here to reset