The Shannon Entropy of a Histogram

10/06/2022
by   Stephen Watts, et al.
0

The histogram is a key method for visualizing data and estimating the underlying probability distribution. Incorrect conclusions about the data result from over or under-binning. A new method based on the Shannon entropy of the histogram uses a simple formula based on the differential entropy estimated from nearest-neighbour distances. Links are made between the new method and other algorithms such as Scott's formula, and cost and risk function methods. A parameter is found that predicts over and under-binning, which can be estimated for any histogram. The new algorithm is shown to be robust by application to real data.

READ FULL TEXT
research
09/24/2018

Shannon Entropy for Neutrosophic Information

The paper presents an extension of Shannon entropy for neutrosophic info...
research
01/13/2023

Learn your entropy from informative data: an axiom ensuring the consistent identification of generalized entropies

Shannon entropy, a cornerstone of information theory, statistical physic...
research
05/09/2012

Improving Compressed Counting

Compressed Counting (CC) [22] was recently proposed for estimating the a...
research
11/14/2019

Estimating differential entropy using recursive copula splitting

A method for estimating the Shannon differential entropy of multidimensi...
research
09/14/2021

Histogram binning revisited with a focus on human perception

This paper presents a quantitative user study to evaluate how well users...
research
04/04/2022

Estimating the Entropy of Linguistic Distributions

Shannon entropy is often a quantity of interest to linguists studying th...
research
09/09/2021

Almost sure convergence of the accelerated weight histogram algorithm

The accelerated weight histogram (AWH) algorithm is an iterative extende...

Please sign up or login with your details

Forgot password? Click here to reset