Exact and Consistent Interpretation for Piecewise Linear Neural Networks: A Closed Form Solution

02/17/2018
by   Lingyang Chu, et al.
0

Strong intelligent machines powered by deep neural networks are increasingly deployed as black boxes to make decisions in risk-sensitive domains, such as finance and medical. To reduce potential risk and build trust with users, it is critical to interpret how such machines make their decisions. Existing works interpret a pre-trained neural network by analyzing hidden neurons, mimicking pre-trained models or approximating local predictions. However, these methods do not provide a guarantee on the exactness and consistency of their interpretation. In this paper, we propose an elegant closed form solution named OpenBox to compute exact and consistent interpretations for the family of Piecewise Linear Neural Networks (PLNN). The major idea is to first transform a PLNN into a mathematically equivalent set of linear classifiers, then interpret each linear classifier by the features that dominate its prediction. We further apply OpenBox to demonstrate the effectiveness of non-negative and sparse constraints on improving the interpretability of PLNNs. The extensive experiments on both synthetic and real world data sets clearly demonstrate the exactness and consistency of our interpretation.

READ FULL TEXT
research
06/17/2019

Exact and Consistent Interpretation of Piecewise Linear Models Hidden behind APIs: A Closed Form Solution

More and more AI services are provided through APIs on cloud where predi...
research
10/21/2019

Semantics for Global and Local Interpretation of Deep Neural Networks

Deep neural networks (DNNs) with high expressiveness have achieved state...
research
12/02/2020

The Self-Simplifying Machine: Exploiting the Structure of Piecewise Linear Neural Networks to Create Interpretable Models

Today, it is more important than ever before for users to have trust in ...
research
03/31/2022

Interpretation of Black Box NLP Models: A Survey

An increasing number of machine learning models have been deployed in do...
research
10/10/2017

LinXGBoost: Extension of XGBoost to Generalized Local Linear Models

XGBoost is often presented as the algorithm that wins every ML competiti...
research
01/28/2022

Linear Adversarial Concept Erasure

Modern neural models trained on textual data rely on pre-trained represe...
research
05/14/2017

Detecting Statistical Interactions from Neural Network Weights

Interpreting deep neural networks can enable new applications for predic...

Please sign up or login with your details

Forgot password? Click here to reset