Learning Optimal Representations with the Decodable Information Bottleneck

09/27/2020
by   Yann Dubois, et al.
16

We address the question of characterizing and finding optimal representations for supervised learning. Traditionally, this question has been tackled using the Information Bottleneck, which compresses the inputs while retaining information about the targets, in a decoder-agnostic fashion. In machine learning, however, our goal is not compression but rather generalization, which is intimately linked to the predictive family or decoder of interest (e.g. linear classifier). We propose the Decodable Information Bottleneck (DIB) that considers information retention and compression from the perspective of the desired predictive family. As a result, DIB gives rise to representations that are optimal in terms of expected test performance and can be estimated with guarantees. Empirically, we show that the framework can be used to enforce a small generalization gap on downstream classifiers and to predict the generalization ability of neural networks.

READ FULL TEXT

page 2

page 6

research
06/06/2019

Class-Conditional Compression and Disentanglement: Bridging the Gap between Neural Networks and Naive Bayes Classifiers

In this draft, which reports on work in progress, we 1) adapt the inform...
research
04/28/2023

Recognizable Information Bottleneck

Information Bottlenecks (IBs) learn representations that generalize to u...
research
03/21/2020

On Information Plane Analyses of Neural Network Classifiers – A Review

We review the current literature concerned with information plane analys...
research
10/25/2022

Characterizing information loss in a chaotic double pendulum with the Information Bottleneck

A hallmark of chaotic dynamics is the loss of information with time. Alt...
research
10/15/2019

Extracting robust and accurate features via a robust information bottleneck

We propose a novel strategy for extracting features in supervised learni...
research
03/05/2019

Improving Generalization of Deep Networks for Inverse Reconstruction of Image Sequences

Deep learning networks have shown state-of-the-art performance in many i...
research
07/28/2021

Characterizing the Generalization Error of Gibbs Algorithm with Symmetrized KL information

Bounding the generalization error of a supervised learning algorithm is ...

Please sign up or login with your details

Forgot password? Click here to reset