Why ReLU networks yield high-confidence predictions far away from the training data and how to mitigate the problem

12/13/2018
by   Matthias Hein, et al.
52

Classifiers used in the wild, in particular for safety-critical systems, should not only have good generalization properties but also should know when they don't know, in particular make low confidence predictions far away from the training data. We show that ReLU type neural networks which yield a piecewise linear classifier function fail in this regard as they produce almost always high confidence predictions far away from the training data. For bounded domains like images we propose a new robust optimization technique similar to adversarial training which enforces low confidence predictions far away from the training data. We show that this technique is surprisingly effective in reducing the confidence of predictions far away from the training data while maintaining high confidence predictions and similar test error on the original classification task compared to standard training.

READ FULL TEXT

page 6

page 11

page 12

page 15

page 16

page 17

research
09/26/2019

Towards neural networks that provably know when they don't know

It has recently been shown that ReLU networks produce arbitrarily over-c...
research
11/17/2021

Do Not Trust Prediction Scores for Membership Inference Attacks

Membership inference attacks (MIAs) aim to determine whether a specific ...
research
10/06/2020

Fixing Asymptotic Uncertainty of Bayesian Neural Networks with Infinite ReLU Features

Approximate Bayesian methods can mitigate overconfidence in ReLU network...
research
04/01/2021

The Compact Support Neural Network

Neural networks are popular and useful in many fields, but they have the...
research
02/24/2020

Being Bayesian, Even Just a Bit, Fixes Overconfidence in ReLU Networks

The point estimates of ReLU classification networks—arguably the most wi...
research
08/30/2022

Constraining Representations Yields Models That Know What They Don't Know

A well-known failure mode of neural networks corresponds to high confide...
research
06/01/2011

Identifying Mislabeled Training Data

This paper presents a new approach to identifying and eliminating mislab...

Please sign up or login with your details

Forgot password? Click here to reset