Mitigating Neural Network Overconfidence with Logit Normalization

05/19/2022
by   Hongxin Wei, et al.
0

Detecting out-of-distribution inputs is critical for safe deployment of machine learning models in the real world. However, neural networks are known to suffer from the overconfidence issue, where they produce abnormally high confidence for both in- and out-of-distribution inputs. In this work, we show that this issue can be mitigated through Logit Normalization (LogitNorm) – a simple fix to the cross-entropy loss – by enforcing a constant vector norm on the logits in training. Our method is motivated by the analysis that the norm of the logit keeps increasing during training, leading to overconfident output. Our key idea behind LogitNorm is thus to decouple the influence of output's norm during network optimization. Trained with LogitNorm, neural networks produce highly distinguishable confidence scores between in- and out-of-distribution data. Extensive experiments demonstrate the superiority of LogitNorm, reducing the average FPR95 by up to 42.30

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/01/2021

On the Importance of Gradients for Detecting Distributional Shifts in the Wild

Detecting out-of-distribution (OOD) data has become a critical component...
research
11/26/2017

Training Confidence-calibrated Classifiers for Detecting Out-of-Distribution Samples

The problem of detecting whether a test sample is from in-distribution (...
research
02/15/2023

Uncertainty-Estimation with Normalized Logits for Out-of-Distribution Detection

Out-of-distribution (OOD) detection is critical for preventing deep lear...
research
05/28/2023

T2FNorm: Extremely Simple Scaled Train-time Feature Normalization for OOD Detection

Neural networks are notorious for being overconfident predictors, posing...
research
11/24/2021

ReAct: Out-of-distribution Detection With Rectified Activations

Out-of-distribution (OOD) detection has received much attention lately d...
research
09/28/2018

Confidence Calibration in Deep Neural Networks through Stochastic Inferences

We propose a generic framework to calibrate accuracy and confidence (sco...
research
11/18/2021

On the Effectiveness of Sparsification for Detecting the Deep Unknowns

Detecting out-of-distribution (OOD) inputs is a central challenge for sa...

Please sign up or login with your details

Forgot password? Click here to reset