Auditing and Debugging Deep Learning Models via Decision Boundaries: Individual-level and Group-level Analysis

01/03/2020
by   Roozbeh Yousefzadeh, et al.
8

Deep learning models have been criticized for their lack of easy interpretation, which undermines confidence in their use for important applications. Nevertheless, they are consistently utilized in many applications, consequential to humans' lives, mostly because of their better performance. Therefore, there is a great need for computational methods that can explain, audit, and debug such models. Here, we use flip points to accomplish these goals for deep learning models with continuous output scores (e.g., computed by softmax), used in social applications. A flip point is any point that lies on the boundary between two output classes: e.g. for a model with a binary yes/no output, a flip point is any input that generates equal scores for "yes" and "no". The flip point closest to a given input is of particular importance because it reveals the least changes in the input that would change a model's classification, and we show that it is the solution to a well-posed optimization problem. Flip points also enable us to systematically study the decision boundaries of a deep learning classifier. The resulting insight into the decision boundaries of a deep model can clearly explain the model's output on the individual-level, via an explanation report that is understandable by non-experts. We also develop a procedure to understand and audit model behavior towards groups of people. Flip points can also be used to alter the decision boundaries in order to improve undesirable behaviors. We demonstrate our methods by investigating several models trained on standard datasets used in social applications of machine learning. We also identify the features that are most responsible for particular classifications and misclassifications.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/21/2019

Interpreting Neural Networks Using Flip Points

Neural networks have been criticized for their lack of easy interpretati...
research
08/07/2019

Investigating Decision Boundaries of Trained Neural Networks

Deep learning models have been the subject of study from various perspec...
research
01/25/2021

Deep Learning Generalization and the Convex Hull of Training Sets

We study the generalization of deep learning models in relation to the c...
research
04/15/2023

The XAISuite framework and the implications of explanatory system dissonance

Explanatory systems make machine learning models more transparent. Howev...
research
02/02/2020

Interpretability of Blackbox Machine Learning Models through Dataview Extraction and Shadow Model creation

Deep learning models trained using massive amounts of data tend to captu...
research
10/01/2019

Sampling Unknown Decision Functions to Build Classifier Copies

Copies have been proposed as a viable alternative to endow machine learn...
research
10/28/2022

I am Only Happy When There is Light: The Impact of Environmental Changes on Affective Facial Expressions Recognition

Human-robot interaction (HRI) benefits greatly from advances in the mach...

Please sign up or login with your details

Forgot password? Click here to reset