Bounds for Learning Lossless Source Coding

09/18/2020
by   Anders Host-Madsen, et al.
0

This paper asks a basic question: how much training is required to beat a universal source coder? Traditionally, there have been two types of source coders: fixed, optimum coders such as Huffman coders; and universal source coders, such as Lempel-Ziv The paper considers a third type of source coders: learned coders. These are coders that are trained on data of a particular type, and then used to encode new data of that type. This is a type of coder that has recently become very popular for (lossy) image and video coding. The paper consider two criteria for performance of learned coders: the average performance over training data, and a guaranteed performance over all training except for some error probability P_e. In both cases the coders are evaluated with respect to redundancy. The paper considers the IID binary case and binary Markov chains. In both cases it is shown that the amount of training data required is very moderate: to code sequences of length l the amount of training data required to beat a universal source coder is m=Kl/log l, where the constant in front depends the case considered.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/18/2018

Optimum Overflow Thresholds in Variable-Length Source Coding Allowing Non-Vanishing Error Probability

The variable-length source coding problem allowing the error probability...
research
11/22/2022

Universal Variable-to-Fixed Length Lossy Compression at Finite Blocklengths

We consider universal variable-to-fixed length compression of memoryless...
research
08/27/2020

Type Size Code for Compressing Erdös-Rényi Graphs

We consider universal source coding of unlabeled graphs which are common...
research
08/20/2018

Optimized Rate-Adaptive Protograph-Based LDPC Codes for Source Coding with Side Information

This paper considers the problem of source coding with side information ...
research
07/11/2021

On Universal D-Semifaithful Coding for Memoryless Sources with Infinite Alphabets

The problem of variable length and fixed-distortion universal source cod...
research
07/20/2021

Single-Shot Compression for Hypothesis Testing

Enhanced processing power in the cloud allows constrained devices to off...
research
08/01/2018

Forest Learning from Data and its Universal Coding

This paper considers structure learning from data with n samples of p va...

Please sign up or login with your details

Forgot password? Click here to reset