Data efficiency, dimensionality reduction, and the generalized symmetric information bottleneck

09/11/2023
by   K. Michael Martini, et al.
0

The Symmetric Information Bottleneck (SIB), an extension of the more familiar Information Bottleneck, is a dimensionality reduction technique that simultaneously compresses two random variables to preserve information between their compressed versions. We introduce the Generalized Symmetric Information Bottleneck (GSIB), which explores different functional forms of the cost of such simultaneous reduction. We then explore the dataset size requirements of such simultaneous compression. We do this by deriving bounds and root-mean-squared estimates of statistical fluctuations of the involved loss functions. We show that, in typical situations, the simultaneous GSIB compression requires qualitatively less data to achieve the same errors compared to compressing variables one at a time. We suggest that this is an example of a more general principle that simultaneous compression is more data efficient than independent compression of each of the input variables.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/16/2020

Difficulty in estimating visual information from randomly sampled images

In this paper, we evaluate dimensionality reduction methods in terms of ...
research
02/09/2022

Reducing Redundancy in the Bottleneck Representation of the Autoencoders

Autoencoders are a type of unsupervised neural networks, which can be us...
research
10/04/2021

Robust Linear Classification from Limited Training Data

We consider the problem of linear classification under general loss func...
research
05/06/2020

Stochastic Bottleneck: Rateless Auto-Encoder for Flexible Dimensionality Reduction

We propose a new concept of rateless auto-encoders (RL-AEs) that enable ...
research
05/18/2023

Information-Ordered Bottlenecks for Adaptive Semantic Compression

We present the information-ordered bottleneck (IOB), a neural layer desi...
research
06/17/2021

On Effects of Compression with Hyperdimensional Computing in Distributed Randomized Neural Networks

A change of the prevalent supervised learning techniques is foreseeable ...
research
02/16/2013

Understanding Boltzmann Machine and Deep Learning via A Confident Information First Principle

Typical dimensionality reduction methods focus on directly reducing the ...

Please sign up or login with your details

Forgot password? Click here to reset