Addressing Randomness in Evaluation Protocols for Out-of-Distribution Detection

03/01/2022
by   Konstantin Kirchheim, et al.
0

Deep Neural Networks for classification behave unpredictably when confronted with inputs not stemming from the training distribution. This motivates out-of-distribution detection (OOD) mechanisms. The usual lack of prior information on out-of-distribution data renders the performance estimation of detection approaches on unseen data difficult. Several contemporary evaluation protocols are based on open set simulations, which average the performance over up to five synthetic random splits of a dataset into in- and out-of-distribution samples. However, the number of possible splits may be much larger, and the performance of Deep Neural Networks is known to fluctuate significantly depending on different sources of random variation. We empirically demonstrate that current protocols may fail to provide reliable estimates of the expected performance of OOD methods. By casting this evaluation as a random process, we generalize the concept of open set simulations and propose to estimate the performance of OOD methods using a Monte Carlo approach that addresses the randomness.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/22/2017

Practical Implementation of a Deep Random Generator

We have introduced in former work the concept of Deep Randomness and its...
research
10/13/2022

Large-Scale Open-Set Classification Protocols for ImageNet

Open-Set Classification (OSC) intends to adapt closed-set classification...
research
07/24/2018

Reliable Uncertainty Estimates in Deep Neural Networks using Noise Contrastive Priors

Obtaining reliable uncertainty estimates of neural network predictions i...
research
07/28/2020

Toward Reliable Models for Authenticating Multimedia Content: Detecting Resampling Artifacts With Bayesian Neural Networks

In multimedia forensics, learning-based methods provide state-of-the-art...
research
02/03/2021

On Entropy and Bit Patterns of Ring Oscillator Jitter

Thermal jitter (phase noise) from a free-running ring oscillator is a co...
research
06/17/2022

Open-Sampling: Exploring Out-of-Distribution data for Re-balancing Long-tailed datasets

Deep neural networks usually perform poorly when the training dataset su...
research
05/20/2019

Inference under Information Constraints II: Communication Constraints and Shared Randomness

A central server needs to perform statistical inference based on samples...

Please sign up or login with your details

Forgot password? Click here to reset