What Information Does a ResNet Compress?

03/13/2020
by   Luke Nicholas Darlow, et al.
4

The information bottleneck principle (Shwartz-Ziv Tishby, 2017) suggests that SGD-based training of deep neural networks results in optimally compressed hidden layers, from an information theoretic perspective. However, this claim was established on toy data. The goal of the work we present here is to test whether the information bottleneck principle is applicable to a realistic setting using a larger and deeper convolutional architecture, a ResNet model. We trained PixelCNN++ models as inverse representation decoders to measure the mutual information between hidden layers of a ResNet and input image data, when trained for (1) classification and (2) autoencoding. We find that two stages of learning happen for both training regimes, and that compression does occur, even for an autoencoder. Sampling images by conditioning on hidden layers' activations offers an intuitive visualisation to understand what a ResNets learns to forget.

READ FULL TEXT

page 1

page 8

page 9

page 12

page 15

page 16

page 17

research
06/24/2021

Information Bottleneck: Exact Analysis of (Quantized) Neural Networks

The information bottleneck (IB) principle has been suggested as a way to...
research
05/13/2023

Information Bottleneck Analysis of Deep Neural Networks via Lossy Compression

The Information Bottleneck (IB) principle offers an information-theoreti...
research
06/13/2020

Understanding Learning Dynamics of Binary Neural Networks via Information Bottleneck

Compact neural networks are essential for affordable and power efficient...
research
10/24/2020

Graph Information Bottleneck

Representation learning of graph-structured data is challenging because ...
research
02/28/2018

Compressing Neural Networks using the Variational Information Bottleneck

Neural networks can be compressed to reduce memory and computational req...
research
03/27/2020

Unpacking Information Bottlenecks: Unifying Information-Theoretic Objectives in Deep Learning

The information bottleneck (IB) principle offers both a mechanism to exp...
research
04/08/2015

A Group Theoretic Perspective on Unsupervised Deep Learning

Why does Deep Learning work? What representations does it capture? How d...

Please sign up or login with your details

Forgot password? Click here to reset