Bayesian estimation of information-theoretic metrics for sparsely sampled distributions

01/31/2023
by   Angelo Piga, et al.
0

Estimating the Shannon entropy of a discrete distribution from which we have only observed a small sample is challenging. Estimating other information-theoretic metrics, such as the Kullback-Leibler divergence between two sparsely sampled discrete distributions, is even harder. Existing approaches to address these problems have shortcomings: they are biased, heuristic, work only for some distributions, and/or cannot be applied to all information-theoretic metrics. Here, we propose a fast, semi-analytical estimator for sparsely sampled distributions that is efficient, precise, and general. Its derivation is grounded in probabilistic considerations and uses a hierarchical Bayesian approach to extract as much information as possible from the few observations available. Our approach provides estimates of the Shannon entropy with precision at least comparable to the state of the art, and most often better. It can also be used to obtain accurate estimates of any other information-theoretic metric, including the notoriously challenging Kullback-Leibler divergence. Here, again, our approach performs consistently better than existing estimators.

READ FULL TEXT
research
07/27/2022

Informational properties of the family of cubic rank transmuted distributions

Recently, cubic rank transmuted (CRT) distribution was introduced and st...
research
07/24/2023

On the information-theoretic formulation of network participation

The participation coefficient is a widely used metric of the diversity o...
research
04/04/2022

Estimating the Entropy of Linguistic Distributions

Shannon entropy is often a quantity of interest to linguists studying th...
research
10/26/2013

Efficient Information Theoretic Clustering on Discrete Lattices

We consider the problem of clustering data that reside on discrete, low ...
research
11/04/2015

Quantification of observed prior and likelihood information in parametric Bayesian modeling

Two data-dependent information metrics are developed to quantify the inf...
research
06/29/2023

Tokenization and the Noiseless Channel

Subword tokenization is a key part of many NLP pipelines. However, littl...
research
12/14/2018

The Entropy of Artificial Intelligence and a Case Study of AlphaZero from Shannon's Perspective

The recently released AlphaZero algorithm achieves superhuman performanc...

Please sign up or login with your details

Forgot password? Click here to reset