Mutual information estimation for graph convolutional neural networks

03/31/2022
by   Marius C. Landverk, et al.
0

Measuring model performance is a key issue for deep learning practitioners. However, we often lack the ability to explain why a specific architecture attains superior predictive accuracy for a given data set. Often, validation accuracy is used as a performance heuristic quantifying how well a network generalizes to unseen data, but it does not capture anything about the information flow in the model. Mutual information can be used as a measure of the quality of internal representations in deep learning models, and the information plane may provide insights into whether the model exploits the available information in the data. The information plane has previously been explored for fully connected neural networks and convolutional architectures. We present an architecture-agnostic method for tracking a network's internal representations during training, which are then used to create the mutual information plane. The method is exemplified for graph-based neural networks fitted on citation data. We compare how the inductive bias introduced in graph-based architectures changes the mutual information plane relative to a fully connected neural network.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/20/2019

NIF: A Framework for Quantifying Neural Information Flow in Deep Networks

In this paper, we present a new approach to interpreting deep learning m...
research
06/01/2018

Inverting Supervised Representations with Autoregressive Neural Density Models

Understanding the nature of representations learned by supervised machin...
research
09/16/2020

Malicious Network Traffic Detection via Deep Learning: An Information Theoretic View

The attention that deep learning has garnered from the academic communit...
research
10/26/2022

InfoShape: Task-Based Neural Data Shaping via Mutual Information

The use of mutual information as a tool in private data sharing has rema...
research
06/10/2020

On the Maximum Mutual Information Capacity of Neural Architectures

We derive the closed-form expression of the maximum mutual information -...
research
01/17/2020

Exact Information Bottleneck with Invertible Neural Networks: Getting the Best of Discriminative and Generative Modeling

The Information Bottleneck (IB) principle offers a unified approach to m...
research
06/27/2022

Monitoring Shortcut Learning using Mutual Information

The failure of deep neural networks to generalize to out-of-distribution...

Please sign up or login with your details

Forgot password? Click here to reset