Inverting Supervised Representations with Autoregressive Neural Density Models

06/01/2018
by   Charlie Nash, et al.
0

Understanding the nature of representations learned by supervised machine learning models is a significant goal in the machine learning community. We present a method for feature interpretation that makes use of recent advances in autoregressive density estimation models to invert model representations. We train generative inversion models to express a distribution over input features conditioned on intermediate model representations. Insights into the invariances learned by supervised models can be gained by viewing samples from these inversion models. In addition, we can use these inversion models to estimate the mutual information between a model's inputs and its intermediate representations, thus quantifying the amount of information preserved by the network at different stages. Using this method we examine the types of information preserved at different layers of convolutional neural networks, and explore the invariances induced by different architectural choices. Finally we show that the mutual information between inputs and network layers decreases over the course of training, supporting recent work by Shwartz-Ziv and Tishby (2017) on the information bottleneck theory of deep learning.

READ FULL TEXT

page 5

page 6

page 7

research
03/31/2022

Mutual information estimation for graph convolutional neural networks

Measuring model performance is a key issue for deep learning practitione...
research
01/30/2018

Transformation Autoregressive Networks

The fundamental task of general density estimation has been of keen inte...
research
03/28/2021

Explaining Representation by Mutual Information

Science is used to discover the law of world. Machine learning can be us...
research
09/16/2020

Malicious Network Traffic Detection via Deep Learning: An Information Theoretic View

The attention that deep learning has garnered from the academic communit...
research
02/18/2020

Conditional Mutual information-based Contrastive Loss for Financial Time Series Forecasting

We present a method for financial time series forecasting using represen...
research
10/01/2021

Do Self-Supervised and Supervised Methods Learn Similar Visual Representations?

Despite the success of a number of recent techniques for visual self-sup...
research
02/20/2018

i-RevNet: Deep Invertible Networks

It is widely believed that the success of deep convolutional networks is...

Please sign up or login with your details

Forgot password? Click here to reset