Regularized Mutual Information Neural Estimation

11/16/2020
by   Kwanghee Choi, et al.
0

With the variational lower bound of mutual information (MI), the estimation of MI can be understood as an optimization task via stochastic gradient descent. In this work, we start by showing how Mutual Information Neural Estimator (MINE) searches for the optimal function T that maximizes the Donsker-Varadhan representation. With our synthetic dataset, we directly observe the neural network outputs during the optimization to investigate why MINE succeeds or fails: We discover the drifting phenomenon, where the constant term of T is shifting through the optimization process, and analyze the instability caused by the interaction between the logsumexp and the insufficient batch size. Next, through theoretical and experimental evidence, we propose a novel lower bound that effectively regularizes the neural network to alleviate the problems of MINE. We also introduce an averaging strategy that produces an unbiased estimate by utilizing multiple batches to mitigate the batch size limitation. Finally, we show that L^2 regularization achieves significant improvements in both discrete and continuous settings.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/08/2019

Data-Efficient Mutual Information Neural Estimator

Measuring Mutual Information (MI) between high-dimensional, continuous, ...
research
07/30/2021

A Training-Based Mutual Information Lower Bound for Large-Scale Systems

We provide a mutual information lower bound that can be used to analyze ...
research
01/12/2018

MINE: Mutual Information Neural Estimation

We argue that the estimation of the mutual information between high dime...
research
02/19/2021

Sequential- and Parallel- Constrained Max-value Entropy Search via Information Lower Bound

Recently, several Bayesian optimization (BO) methods have been extended ...
research
05/10/2021

Neural Computation of Capacity Region of Memoryless Multiple Access Channels

This paper provides a numerical framework for computing the achievable r...
research
03/28/2023

Information-Theoretic GAN Compression with Variational Energy-based Model

We propose an information-theoretic knowledge distillation approach for ...
research
02/15/2021

Posterior-Aided Regularization for Likelihood-Free Inference

The recent development of likelihood-free inference aims training a flex...

Please sign up or login with your details

Forgot password? Click here to reset