Prediction by Compression

08/30/2010
by   Joel Ratsaby, et al.
0

It is well known that text compression can be achieved by predicting the next symbol in the stream of text data based on the history seen up to the current symbol. The better the prediction the more skewed the conditional probability distribution of the next symbol and the shorter the codeword that needs to be assigned to represent this next symbol. What about the opposite direction ? suppose we have a black box that can compress text stream. Can it be used to predict the next symbol in the stream ? We introduce a criterion based on the length of the compressed data and use it to predict the next symbol. We examine empirically the prediction error rate and its dependency on some compression parameters.

READ FULL TEXT

page 7

page 8

page 9

page 10

research
09/06/2022

Compression Optimality of Asymmetric Numeral Systems

Compression also known as entropy coding has a rich and long history. Ho...
research
10/02/2021

Complete b-symbol weight distribution of some irreducible cyclic codes

Recently, b-symbol codes are proposed to protect against b-symbol errors...
research
04/30/2010

Graphic Symbol Recognition using Graph Based Signature and Bayesian Network Classifier

We present a new approach for recognition of complex graphic symbols in ...
research
05/07/2022

Rate-Constrained Shaping Codes for Finite-State Channels With Cost

Shaping codes are used to generate code sequences in which the symbols o...
research
08/14/2023

Ergodic Estimations for Toeplitz Sequences Generated by a Symbol

We analyse the convergence of the ergodic formula for Toeplitz matrix-se...
research
07/01/2021

Data Deduplication with Random Substitutions

Data deduplication saves storage space by identifying and removing repea...
research
05/03/2019

Preservation of normality by non-oblivious group selection

We give two different proofs of the fact that non-oblivious selection vi...

Please sign up or login with your details

Forgot password? Click here to reset