Why Calibration Error is Wrong Given Model Uncertainty: Using Posterior Predictive Checks with Deep Learning

12/02/2021
by   Achintya Gopal, et al.
0

Within the last few years, there has been a move towards using statistical models in conjunction with neural networks with the end goal of being able to better answer the question, "what do our models know?". From this trend, classical metrics such as Prediction Interval Coverage Probability (PICP) and new metrics such as calibration error have entered the general repertoire of model evaluation in order to gain better insight into how the uncertainty of our model compares to reality. One important component of uncertainty modeling is model uncertainty (epistemic uncertainty), a measurement of what the model does and does not know. However, current evaluation techniques tends to conflate model uncertainty with aleatoric uncertainty (irreducible error), leading to incorrect conclusions. In this paper, using posterior predictive checks, we show how calibration error and its variants are almost always incorrect to use given model uncertainty, and further show how this mistake can lead to trust in bad models and mistrust in good models. Though posterior predictive checks has often been used for in-sample evaluation of Bayesian models, we show it still has an important place in the modern deep learning world.

READ FULL TEXT

page 9

page 22

research
08/26/2019

Open Set Recognition Through Deep Neural Network Uncertainty: Does Out-of-Distribution Detection Require Generative Classifiers?

We present an analysis of predictive uncertainty based out-of-distributi...
research
10/28/2019

Towards calibrated and scalable uncertainty representations for neural networks

For many applications it is critical to know the uncertainty of a neural...
research
07/17/2022

Uncertainty Calibration in Bayesian Neural Networks via Distance-Aware Priors

As we move away from the data, the predictive uncertainty should increas...
research
06/04/2019

Revision of ISO 19229 to support the certification of calibration gases for purity

The second edition of ISO 19229 expands the guidance in its predecessor ...
research
11/23/2021

Weight Pruning and Uncertainty in Radio Galaxy Classification

In this work we use variational inference to quantify the degree of epis...
research
05/24/2016

Posterior Dispersion Indices

Probabilistic modeling is cyclical: we specify a model, infer its poster...
research
03/21/2017

Overcoming model simplifications when quantifying predictive uncertainty

It is generally accepted that all models are wrong -- the difficulty is ...

Please sign up or login with your details

Forgot password? Click here to reset