Understanding Convolutional Neural Network Training with Information Theory

04/18/2018
by   Shujian Yu, et al.
0

Using information theoretic concepts to understand and explore the inner organization of deep neural networks (DNNs) remains a big challenge. Recently, the concept of an information plane began to shed light on the analysis of multilayer perceptrons (MLPs). We provided an in-depth insight into stacked autoencoders (SAEs) using a novel matrix-based Renyi's α-entropy functional, enabling for the first time the analysis of the dynamics of learning using information flow in real-world scenario involving complex network architecture and large data. Despite the great potential of these past works, there are several open questions when it comes to applying information theoretic concepts to understand convolutional neural networks (CNNs). These include for instance the accurate estimation of information quantities among multiple variables, and the many different training methodologies. By extending the novel matrix-based Renyi's α-entropy functional to a multivariate scenario, this paper presents a systematic method to analyze CNNs training using information theory. Our results validate two fundamental data processing inequalities in CNNs, and also have direct impacts on previous work concerning the training and design of CNNs.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/30/2018

Understanding Autoencoders with Information Theoretic Concepts

Despite their great success in practical applications, there is still a ...
research
05/02/2020

An Information-theoretic Visual Analysis Framework for Convolutional Neural Networks

Despite the great success of Convolutional Neural Networks (CNNs) in Com...
research
09/25/2019

Information Plane Analysis of Deep Neural Networks via Matrix-Based Renyi's Entropy and Tensor Kernels

Analyzing deep neural networks (DNNs) via information plane (IP) theory ...
research
11/09/2019

Information Bottleneck Methods on Convolutional Neural Networks

Recent year, many researches attempt to open the black box of deep neura...
research
05/21/2020

A Neural Network Looks at Leonardo's(?) Salvator Mundi

We use convolutional neural networks (CNNs) to analyze authorship questi...
research
11/19/2018

Deeper Interpretability of Deep Networks

Deep Convolutional Neural Networks (CNNs) have been one of the most infl...
research
02/05/2021

Learning While Dissipating Information: Understanding the Generalization Capability of SGLD

Understanding the generalization capability of learning algorithms is at...

Please sign up or login with your details

Forgot password? Click here to reset