On the Difference Between the Information Bottleneck and the Deep Information Bottleneck

12/31/2019
by   Aleksander Wieczorek, et al.
30

Combining the Information Bottleneck model with deep learning by replacing mutual information terms with deep neural nets has proved successful in areas ranging from generative modelling to interpreting deep neural networks. In this paper, we revisit the Deep Variational Information Bottleneck and the assumptions needed for its derivation. The two assumed properties of the data X, Y and their latent representation T take the form of two Markov chains T-X-Y and X-T-Y. Requiring both to hold during the optimisation process can be limiting for the set of potential joint distributions P(X,Y,T). We therefore show how to circumvent this limitation by optimising a lower bound for I(T;Y) for which only the latter Markov chain has to be satisfied. The actual mutual information consists of the lower bound which is optimised in DVIB and cognate models in practice and of two terms measuring how much the former requirement T-X-Y is violated. Finally, we propose to interpret the family of information bottleneck models as directed graphical models and show that in this framework the original and deep information bottlenecks are special cases of a fundamental IB model.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/02/2019

Information bottleneck through variational glasses

Information bottleneck (IB) principle [1] has become an important elemen...
research
11/12/2019

Learning Representations in Reinforcement Learning:An Information Bottleneck Approach

The information bottleneck principle is an elegant and useful approach t...
research
08/22/2022

On Information Bottleneck for Gaussian Processes

The information bottleneck problem (IB) of jointly stationary Gaussian s...
research
05/09/2022

The Compound Information Bottleneck Outlook

We formulate and analyze the compound information bottleneck programming...
research
09/29/2015

Variational Information Maximisation for Intrinsically Motivated Reinforcement Learning

The mutual information is a core statistical quantity that has applicati...
research
12/04/2022

Statistical Physics of Deep Neural Networks: Initialization toward Optimal Channels

In deep learning, neural networks serve as noisy channels between input ...
research
06/08/2020

The Dual Information Bottleneck

The Information Bottleneck (IB) framework is a general characterization ...

Please sign up or login with your details

Forgot password? Click here to reset