Understanding Autoencoders with Information Theoretic Concepts

03/30/2018
by   Shujian Yu, et al.
0

Despite their great success in practical applications, there is still a lack of theoretical and systematic methods to analyze deep neural networks. In this paper, we illustrate an advanced information theoretic methodology to understand the dynamics of learning and the design of autoencoders, a special type of deep learning architectures that resembles a communication channel. By generalizing the information plane to any cost function, and inspecting the roles and dynamics of different layers using layer-wise information quantities, we emphasize the role that mutual information plays in quantifying learning from data. We further propose and also experimentally validate, for mean square error training, two hypotheses regarding the layer-wise flow of information and intrinsic dimensionality of the bottleneck layer, using respectively the data processing inequality and the identification of a bifurcation point in the information plane that is controlled by the given data. Our observations have direct impact on the optimal design of autoencoders, the design of alternative feedforward training methods, and even in the problem of generalization.

READ FULL TEXT

page 7

page 17

page 20

research
04/18/2018

Understanding Convolutional Neural Network Training with Information Theory

Using information theoretic concepts to understand and explore the inner...
research
05/15/2020

On the Information Plane of Autoencoders

The training dynamics of hidden layers in deep learning are poorly under...
research
02/21/2018

Information Theoretic Co-Training

This paper introduces an information theoretic co-training objective for...
research
10/20/2019

Towards Further Understanding of Sparse Filtering via Information Bottleneck

In this paper we examine a formalization of feature distribution learnin...
research
01/17/2020

Exact Information Bottleneck with Invertible Neural Networks: Getting the Best of Discriminative and Generative Modeling

The Information Bottleneck (IB) principle offers a unified approach to m...
research
04/30/2020

The Information Bottleneck Problem and Its Applications in Machine Learning

Inference capabilities of machine learning (ML) systems skyrocketed in r...
research
01/27/2021

Variational Encoders and Autoencoders : Information-theoretic Inference and Closed-form Solutions

This work develops problem statements related to encoders and autoencode...

Please sign up or login with your details

Forgot password? Click here to reset