Take 5: Interpretable Image Classification with a Handful of Features

03/23/2023
by   Thomas Norrenbrock, et al.
0

Deep Neural Networks use thousands of mostly incomprehensible features to identify a single class, a decision no human can follow. We propose an interpretable sparse and low dimensional final decision layer in a deep neural network with measurable aspects of interpretability and demonstrate it on fine-grained image classification. We argue that a human can only understand the decision of a machine learning model, if the features are interpretable and only very few of them are used for a single decision. For that matter, the final layer has to be sparse and, to make interpreting the features feasible, low dimensional. We call a model with a Sparse Low-Dimensional Decision SLDD-Model. We show that a SLDD-Model is easier to interpret locally and globally than a dense high-dimensional decision layer while being able to maintain competitive accuracy. Additionally, we propose a loss function that improves a model's feature diversity and accuracy. Our more interpretable SLDD-Model only uses 5 out of just 50 features per class, while maintaining 97 to 100 baseline model with 2048 features.

READ FULL TEXT

page 7

page 8

page 16

page 19

page 20

page 21

page 22

page 23

research
12/14/2021

Identifying Class Specific Filters with L1 Norm Frequency Histograms in Deep CNNs

Interpretability of Deep Neural Networks has become a major area of expl...
research
08/20/2019

TabNet: Attentive Interpretable Tabular Learning

We propose a novel high-performance interpretable deep tabular data lear...
research
05/31/2018

Interpretable Set Functions

We propose learning flexible but interpretable functions that aggregate ...
research
02/05/2021

Interpretable Neural Networks based classifiers for categorical inputs

Because of the pervasive usage of Neural Networks in human sensitive app...
research
07/21/2016

Explaining Classification Models Built on High-Dimensional Sparse Data

Predictive modeling applications increasingly use data representing peop...
research
07/24/2023

Feature Gradient Flow for Interpreting Deep Neural Networks in Head and Neck Cancer Prediction

This paper introduces feature gradient flow, a new technique for interpr...
research
07/08/2023

Improving Prototypical Part Networks with Reward Reweighing, Reselection, and Retraining

In recent years, work has gone into developing deep interpretable method...

Please sign up or login with your details

Forgot password? Click here to reset