Entropy estimation of symbol sequences

03/21/2002
by   Thomas Schürmann, et al.
0

We discuss algorithms for estimating the Shannon entropy h of finite symbol sequences with long range correlations. In particular, we consider algorithms which estimate h from the code lengths produced by some compression algorithm. Our interest is in describing their convergence with sequence length, assuming no limits for the space and time complexities of the compression algorithms. A scaling law is proposed for extrapolation from finite sample lengths. This is applied to sequences of dynamical systems in non-trivial chaotic regimes, a 1-D cellular automaton, and to written English texts.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/12/2016

Factorization of C-finite Sequences

We discuss how to decide whether a given C-finite sequence can be writte...
research
01/03/2014

Computing Entropy Rate Of Symbol Sources & A Distribution-free Limit Theorem

Entropy rate of sequential data-streams naturally quantifies the complex...
research
05/03/2019

Preservation of normality by non-oblivious group selection

We give two different proofs of the fact that non-oblivious selection vi...
research
07/05/2021

Time-irreversibility tests for random-length time series: the matching-time approach applied to DNA

In this work we implement the so-called matching time estimators for est...
research
05/05/2022

Cut-Down de Bruijn Sequences

A cut-down de Bruijn sequence is a cyclic string of length L, where 1 ≤ ...
research
07/07/2018

A Note on the Shannon Entropy of Short Sequences

For source sequences of length L symbols we proposed to use a more reali...
research
05/23/2018

Determining the Number of Samples Required to Estimate Entropy in Natural Sequences

Calculating the Shannon entropy for symbolic sequences has been widely c...

Please sign up or login with your details

Forgot password? Click here to reset