A Simple Nadaraya-Watson Head can offer Explainable and Calibrated Classification

12/07/2022
by   Alan Q. Wang, et al.
0

In this paper, we empirically analyze a simple, non-learnable, and nonparametric Nadaraya-Watson (NW) prediction head that can be used with any neural network architecture. In the NW head, the prediction is a weighted average of labels from a support set. The weights are computed from distances between the query feature and support features. This is in contrast to the dominant approach of using a learnable classification head (e.g., a fully-connected layer) on the features, which can be challenging to interpret and can yield poorly calibrated predictions. Our empirical results on an array of computer vision tasks demonstrate that the NW head can yield better calibration than its parametric counterpart, while having comparable accuracy and with minimal computational overhead. To further increase inference-time efficiency, we propose a simple approach that involves a clustering step run on the training set to create a relatively small distilled support set. In addition to using the weights as a means of interpreting model predictions, we further present an easy-to-compute "support influence function," which quantifies the influence of a support element on the prediction for a given query. As we demonstrate in our experiments, the influence function can allow the user to debug a trained model. We believe that the NW head is a flexible, interpretable, and highly useful building block that can be used in a range of applications.

READ FULL TEXT

page 16

page 17

page 18

page 19

page 20

page 22

page 23

page 24

research
04/13/2019

Rethinking Classification and Localization in R-CNN

Modern R-CNN based detectors share the RoI feature extractor head for bo...
research
03/02/2023

Multi-Head Multi-Loss Model Calibration

Delivering meaningful uncertainty estimates is essential for a successfu...
research
12/27/2022

Annealing Double-Head: An Architecture for Online Calibration of Deep Neural Networks

Model calibration, which is concerned with how frequently the model pred...
research
05/24/2020

Deep Convolutional Neural Network-based Bernoulli Heatmap for Head Pose Estimation

Head pose estimation is a crucial problem for many tasks, such as driver...
research
12/22/2017

Obtaining Accurate Probabilistic Causal Inference by Post-Processing Calibration

Discovery of an accurate causal Bayesian network structure from observat...
research
07/24/2023

Feature Gradient Flow for Interpreting Deep Neural Networks in Head and Neck Cancer Prediction

This paper introduces feature gradient flow, a new technique for interpr...
research
01/20/2023

Open-Set Likelihood Maximization for Few-Shot Learning

We tackle the Few-Shot Open-Set Recognition (FSOSR) problem, i.e. classi...

Please sign up or login with your details

Forgot password? Click here to reset