A Law of Data Separation in Deep Learning

10/31/2022
by   Hangfeng He, et al.
0

Multilayer neural networks have achieved superhuman performance in many artificial intelligence applications. However, their black-box nature obscures the underlying mechanism for transforming input data into labels throughout all layers, thus hindering architecture design for new tasks and interpretation for high-stakes decision makings. We addressed this problem by introducing a precise law that governs how real-world deep neural networks separate data according to their class membership from the bottom layers to the top layers in classification problems. This law shows that each layer roughly improves a certain measure of data separation by an equal multiplicative factor. This law manifests in modern architectures such as AlexNet, VGGNet, and ResNet in the late phase of training. This law together with the perspective of data separation offers practical guidelines for designing network architectures, improving model robustness and out-of-sample performance during training, as well as interpreting deep learning predictions.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/04/2018

Power Law in Sparsified Deep Neural Networks

The power law has been observed in the degree distributions of many biol...
research
06/12/2019

Evaluation of Dataflow through layers of Deep Neural Networks in Classification and Regression Problems

This paper introduces two straightforward, effective indices to evaluate...
research
07/09/2019

Characterizing Inter-Layer Functional Mappings of Deep Learning Models

Deep learning architectures have demonstrated state-of-the-art performan...
research
04/15/2022

The training response law explains how deep neural networks learn

Deep neural network is the widely applied technology in this decade. In ...
research
11/15/2022

Power-law Scaling to Assist with Key Challenges in Artificial Intelligence

Power-law scaling, a central concept in critical phenomena, is found to ...
research
07/12/2020

Locality Guided Neural Networks for Explainable Artificial Intelligence

In current deep network architectures, deeper layers in networks tend to...
research
11/15/2022

Efficient shallow learning as an alternative to deep learning

The realization of complex classification tasks requires training of dee...

Please sign up or login with your details

Forgot password? Click here to reset