Estimating Information-Theoretic Quantities with Random Forests

06/30/2019
by   Richard Guo, et al.
1

Information-theoretic quantities, such as mutual information and conditional entropy, are useful statistics for measuring the dependence between two random variables. However, estimating these quantities in a non-parametric fashion is difficult, especially when the variables are high-dimensional, a mixture of continuous and discrete values, or both. In this paper, we propose a decision forest method, Conditional Forests (CF), to estimate these quantities. By combining quantile regression forests with honest sampling, and introducing a finite sample correction, CF improves finite sample bias in a range of settings. We demonstrate through simulations that CF achieves smaller bias and variance in both low- and high-dimensional settings for estimating posteriors, conditional entropy, and mutual information. We then use CF to estimate the amount of information between neuron class and other ceulluar feautres.

READ FULL TEXT

page 1

page 8

research
06/14/2017

Information Potential Auto-Encoders

In this paper, we suggest a framework to make use of mutual information ...
research
12/19/2014

Information-Theoretic Methods for Identifying Relationships among Climate Variables

Information-theoretic quantities, such as entropy, are used to quantify ...
research
01/27/2021

The Most Informative Order Statistic and its Application to Image Denoising

We consider the problem of finding the subset of order statistics that c...
research
02/24/2022

Estimators of Entropy and Information via Inference in Probabilistic Models

Estimating information-theoretic quantities such as entropy and mutual i...
research
10/31/2022

A robust estimator of mutual information for deep learning interpretability

We develop the use of mutual information (MI), a well-established metric...
research
01/13/2021

Estimating Conditional Mutual Information for Discrete-Continuous Mixtures using Multi-Dimensional Adaptive Histograms

Estimating conditional mutual information (CMI) is an essential yet chal...
research
10/27/2020

A Probabilistic Representation of Deep Learning for Improving The Information Theoretic Interpretability

In this paper, we propose a probabilistic representation of MultiLayer P...

Please sign up or login with your details

Forgot password? Click here to reset