Disentangling deep neural networks with rectified linear units using duality

Despite their success deep neural networks (DNNs) are still largely considered as black boxes. The main issue is that the linear and non-linear operations are entangled in every layer, making it hard to interpret the hidden layer outputs. In this paper, we look at DNNs with rectified linear units (ReLUs), and focus on the gating property (`on/off' states) of the ReLUs. We extend the recently developed dual view in which the computation is broken path-wise to show that learning in the gates is more crucial, and learning the weights given the gates is characterised analytically via the so called neural path kernel (NPK) which depends on inputs and gates. In this paper, we present novel results to show that convolution with global pooling and skip connection provide respectively rotational invariance and ensemble structure to the NPK. To address `black box'-ness, we propose a novel interpretable counterpart of DNNs with ReLUs namely deep linearly gated networks (DLGN): the pre-activations to the gates are generated by a deep linear network, and the gates are then applied as external masks to learn the weights in a different network. The DLGN is not an alternative architecture per se, but a disentanglement and an interpretable re-arrangement of the computations in a DNN with ReLUs. The DLGN disentangles the computations into two `mathematically' interpretable linearities (i) the `primal' linearity between the input and the pre-activations in the gating network and (ii) the `dual' linearity in the path space in the weights network characterised by the NPK. We compare the performance of DNN, DGN and DLGN on CIFAR-10 and CIFAR-100 to show that, the DLGN recovers more than 83.5% of the performance of state-of-the-art DNNs. This brings us to an interesting question: `Is DLGN a universal spectral approximator?'

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/24/2021

Towards Fully Interpretable Deep Neural Networks: Are We There Yet?

Despite the remarkable performance, Deep Neural Networks (DNNs) behave a...
research
09/25/2017

Generative learning for deep networks

Learning, taking into account full distribution of the data, referred to...
research
06/01/2021

More Behind Your Electricity Bill: a Dual-DNN Approach to Non-Intrusive Load Monitoring

Non-intrusive load monitoring (NILM) is a well-known single-channel blin...
research
08/05/2019

NeuroMask: Explaining Predictions of Deep Neural Networks through Mask Learning

Deep Neural Networks (DNNs) deliver state-of-the-art performance in many...
research
03/25/2020

Deep Networks as Logical Circuits: Generalization and Interpretation

Not only are Deep Neural Networks (DNNs) black box models, but also we f...
research
12/01/2015

Analyzing Classifiers: Fisher Vectors and Deep Neural Networks

Fisher Vector classifiers and Deep Neural Networks (DNNs) are popular an...
research
06/26/2020

Interpretable Factorization for Neural Network ECG Models

The ability of deep learning (DL) to improve the practice of medicine an...

Please sign up or login with your details

Forgot password? Click here to reset