Outside the Box: Abstraction-Based Monitoring of Neural Networks

11/20/2019
by   Thomas A. Henzinger, et al.
0

Neural networks have demonstrated unmatched performance in a range of classification tasks. Despite numerous efforts of the research community, novelty detection remains one of the significant limitations of neural networks. The ability to identify previously unseen inputs as novel is crucial for our understanding of the decisions made by neural networks. At runtime, inputs not falling into any of the categories learned during training cannot be classified correctly by the neural network. Existing approaches treat the neural network as a black box and try to detect novel inputs based on the confidence of the output predictions. However, neural networks are not trained to reduce their confidence for novel inputs, which limits the effectiveness of these approaches. We propose a framework to monitor a neural network by observing the hidden layers. We employ a common abstraction from program analysis - boxes - to identify novel behaviors in the monitored layers, i.e., inputs that cause behaviors outside the box. For each neuron, the boxes range over the values seen in training. The framework is efficient and flexible to achieve a desired trade-off between raising false warnings and detecting novel inputs. We illustrate the performance and the robustness to variability in the unknown classes on popular image-classification benchmarks.

READ FULL TEXT

page 6

page 7

research
12/29/2022

Detection of out-of-distribution samples using binary neuron activation patterns

Deep neural networks (DNN) have outstanding performance in various appli...
research
04/25/2021

Customizable Reference Runtime Monitoring of Neural Networks using Resolution Boxes

We present an approach for monitoring classification systems via data ab...
research
09/14/2020

Into the unknown: Active monitoring of neural networks

Machine-learning techniques achieve excellent performance in modern appl...
research
09/06/2018

Adversarial Reprogramming of Sequence Classification Neural Networks

Adversarial Reprogramming has demonstrated success in utilizing pre-trai...
research
09/10/2020

Improved Robustness to Open Set Inputs via Tempered Mixup

Supervised classification methods often assume that evaluation data is d...
research
09/14/2017

Denoising Autoencoders for Overgeneralization in Neural Networks

Despite the recent developments that allowed neural networks to achieve ...
research
08/05/2022

FBI: Fingerprinting models with Benign Inputs

Recent advances in the fingerprinting of deep neural networks detect ins...

Please sign up or login with your details

Forgot password? Click here to reset