G-Distillation: Reducing Overconfident Errors on Novel Samples

04/09/2018
by   Zhizhong Li, et al.
0

Counter to the intuition that unfamiliarity should lead to lack of confidence, current algorithms often make highly confident yet wrong predictions when faced with unexpected test samples from an unknown distribution different from training. Unlike all domain adaptation methods, we cannot gather an "unexpected dataset" prior to test. We propose a simple solution that reduces overconfident errors of samples from an unknown novel distribution without increasing evaluation time: train an ensemble of classifiers and then distill into a single model using both labeled and unlabeled examples. Experimentally, we investigate the overconfidence problem and evaluate our solution by creating "familiar" and "novel" test splits, where "familiar" are identically distributed with training and "novel" are not. We show that our solution yields more appropriate prediction confidences, on familiar and novel data, compared to single models and ensembles distilled on training data only. For example, we reduce confident errors in gender recognition by 94

READ FULL TEXT

page 2

page 5

page 19

research
06/18/2020

Fully Test-time Adaptation by Entropy Minimization

Faced with new and different data during testing, a model must adapt its...
research
04/21/2023

Towards Realizing the Value of Labeled Target Samples: a Two-Stage Approach for Semi-Supervised Domain Adaptation

Semi-Supervised Domain Adaptation (SSDA) is a recently emerging research...
research
10/27/2021

Diversity Matters When Learning From Ensembles

Deep ensembles excel in large-scale image classification tasks both in t...
research
01/27/2020

A Primer on Domain Adaptation

Standard supervised machine learning assumes that the distribution of th...
research
03/05/2017

A Theory of Output-Side Unsupervised Domain Adaptation

When learning a mapping from an input space to an output space, the assu...
research
03/02/2023

Do Machine Learning Models Learn Common Sense?

Machine learning models can make basic errors that are easily hidden wit...
research
02/25/2021

Generalized Adversarial Distances to Efficiently Discover Classifier Errors

Given a black-box classification model and an unlabeled evaluation datas...

Please sign up or login with your details

Forgot password? Click here to reset