Characterizing Inter-Layer Functional Mappings of Deep Learning Models

07/09/2019
by   Donald Waagen, et al.
1

Deep learning architectures have demonstrated state-of-the-art performance for object classification and have become ubiquitous in commercial products. These methods are often applied without understanding (a) the difficulty of a classification task given the input data, and (b) how a specific deep learning architecture transforms that data. To answer (a) and (b), we illustrate the utility of a multivariate nonparametric estimator of class separation, the Henze-Penrose (HP) statistic, in the original as well as layer-induced representations. Given an N-class problem, our contribution defines the C(N,2) combinations of HP statistics as a sample from a distribution of class-pair separations. This allows us to characterize the distributional change to class separation induced at each layer of the model. Fisher permutation tests are used to detect statistically significant changes within a model. By comparing the HP statistic distributions between layers, one can statistically characterize: layer adaptation during training, the contribution of each layer to the classification task, and the presence or absence of consistency between training and validation data. This is demonstrated for a simple deep neural network using CIFAR10 with random-labels, CIFAR10, and MNIST datasets.

READ FULL TEXT

page 6

page 14

page 20

page 21

page 22

page 23

page 24

page 37

research
11/05/2018

How deep is deep enough? - Optimizing deep neural network architecture

Deep neural networks use stacked layers of feature detectors to repeated...
research
10/31/2022

A Law of Data Separation in Deep Learning

Multilayer neural networks have achieved superhuman performance in many ...
research
10/11/2020

Is It Time to Redefine the Classification Task for Deep Neural Networks?

Deep neural networks (DNNs) is demonstrated to be vulnerable to the adve...
research
02/10/2022

Decomposing neural networks as mappings of correlation functions

Understanding the functional principles of information processing in dee...
research
09/01/2022

Complexity of Representations in Deep Learning

Deep neural networks use multiple layers of functions to map an object r...
research
06/21/2019

Intermediate efficiency of some weighted goodness-of-fit statistics

This paper compares the Anderson-Darling and some Eicker-Jaeschke statis...
research
07/12/2021

Improving the Algorithm of Deep Learning with Differential Privacy

In this paper, an adjustment to the original differentially private stoc...

Please sign up or login with your details

Forgot password? Click here to reset