PAC-Bayes Information Bottleneck

09/29/2021
by   Zifeng Wang, et al.
0

Information bottleneck (IB) depicts a trade-off between the accuracy and conciseness of encoded representations. IB has succeeded in explaining the objective and behavior of neural networks (NNs) as well as learning better representations. However, there are still critics of the universality of IB, e.g., phase transition usually fades away, representation compression is not causally related to generalization, and IB is trivial in deterministic cases. In this work, we build a new IB based on the trade-off between the accuracy and complexity of learned weights of NNs. We argue that this new IB represents a more solid connection to the objective of NNs since the information stored in weights (IIW) bounds their PAC-Bayes generalization capability, hence we name it as PAC-Bayes IB (PIB). On IIW, we can identify the phase transition phenomenon in general cases and solidify the causality between compression and generalization. We then derive a tractable solution of PIB and design a stochastic inference algorithm by Markov chain Monte Carlo sampling. We empirically verify our claims through extensive experiments. We also substantiate the superiority of the proposed algorithm on training NNs.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/28/2023

Recognizable Information Bottleneck

Information Bottlenecks (IBs) learn representations that generalize to u...
research
06/17/2021

PAC-Bayes, MAC-Bayes and Conditional Mutual Information: Fast rate bounds that handle general VC classes

We give a novel, unified derivation of conditional PAC-Bayesian and mutu...
research
05/29/2019

Where is the Information in a Deep Neural Network?

Whatever information a Deep Neural Network has gleaned from past data is...
research
06/05/2017

Emergence of Invariance and Disentangling in Deep Representations

Using established principles from Information Theory and Statistics, we ...
research
07/17/2019

Learnability for the Information Bottleneck

The Information Bottleneck (IB) method (tishby2000information) provides ...
research
06/06/2019

Class-Conditional Compression and Disentanglement: Bridging the Gap between Neural Networks and Naive Bayes Classifiers

In this draft, which reports on work in progress, we 1) adapt the inform...
research
08/17/2018

Multiview Boosting by Controlling the Diversity and the Accuracy of View-specific Voters

In this paper we propose a boosting based multiview learning algorithm, ...

Please sign up or login with your details

Forgot password? Click here to reset