Feature Gradient Flow for Interpreting Deep Neural Networks in Head and Neck Cancer Prediction

07/24/2023
by   Yinzhu Jin, et al.
0

This paper introduces feature gradient flow, a new technique for interpreting deep learning models in terms of features that are understandable to humans. The gradient flow of a model locally defines nonlinear coordinates in the input data space representing the information the model is using to make its decisions. Our idea is to measure the agreement of interpretable features with the gradient flow of a model. To then evaluate the importance of a particular feature to the model, we compare that feature's gradient flow measure versus that of a baseline noise feature. We then develop a technique for training neural networks to be more interpretable by adding a regularization term to the loss function that encourages the model gradients to align with those of chosen interpretable features. We test our method in a convolutional neural network prediction of distant metastasis of head and neck cancer from a computed tomography dataset from the Cancer Imaging Archive.

READ FULL TEXT
research
07/06/2023

Generalizing Backpropagation for Gradient-Based Interpretability

Many popular feature-attribution methods for interpreting deep neural ne...
research
03/23/2023

Take 5: Interpretable Image Classification with a Handful of Features

Deep Neural Networks use thousands of mostly incomprehensible features t...
research
02/21/2020

An Investigation of Interpretability Techniques for Deep Learning in Predictive Process Analytics

This paper explores interpretability techniques for two of the most succ...
research
01/20/2019

NIF: A Framework for Quantifying Neural Information Flow in Deep Networks

In this paper, we present a new approach to interpreting deep learning m...
research
02/02/2022

MPVNN: Mutated Pathway Visible Neural Network Architecture for Interpretable Prediction of Cancer-specific Survival Risk

Survival risk prediction using gene expression data is important in maki...
research
09/08/2018

Interpreting Neural Networks With Nearest Neighbors

Local model interpretation methods explain individual predictions by ass...
research
12/07/2022

A Simple Nadaraya-Watson Head can offer Explainable and Calibrated Classification

In this paper, we empirically analyze a simple, non-learnable, and nonpa...

Please sign up or login with your details

Forgot password? Click here to reset