Confidence Scoring Using Whitebox Meta-models with Linear Classifier Probes

05/14/2018
by   Tongfei Chen, et al.
0

We propose a confidence scoring mechanism for multi-layer neural networks based on a paradigm of a base model and a meta-model. The confidence score is learned by the meta-model using features derived from the base model -- a deep multi-layer neural network -- considered a whitebox. As features, we investigate linear classifier probes inserted between the various layers of the base model and trained using each layer's intermediate activations. Experiments show that this approach outperforms various baselines in a filtering task, i.e., task of rejecting samples with low confidence. Experimental results are presented using CIFAR-10 and CIFAR-100 dataset with and without added noise exploring various aspects of the method.

READ FULL TEXT
research
03/19/2021

MetaLabelNet: Learning to Generate Soft-Labels from Noisy-Labels

Real-world datasets commonly have noisy labels, which negatively affects...
research
05/22/2017

Regularizing deep networks using efficient layerwise adversarial training

Adversarial training has been shown to regularize deep neural networks i...
research
04/15/2019

Multi-Head Multi-Layer Attention to Deep Language Representations for Grammatical Error Detection

It is known that a deep neural network model pre-trained with large-scal...
research
07/19/2018

Improving Simple Models with Confidence Profiles

In this paper, we propose a new method called ProfWeight for transferrin...
research
02/27/2020

Transductive Few-shot Learning with Meta-Learned Confidence

We propose a novel transductive inference framework for metric-based met...
research
10/05/2016

Understanding intermediate layers using linear classifier probes

Neural network models have a reputation for being black boxes. We propos...

Please sign up or login with your details

Forgot password? Click here to reset