The Dispersion of the Gauss-Markov Source

04/25/2018
by   Peida Tian, et al.
0

The Gauss-Markov source produces U_i = aU_i-1 + Z_i for i≥ 1, where U_0 = 0, |a|<1 and Z_i∼N(0, σ^2) are i.i.d. Gaussian random variables. We consider lossy compression of a block of n samples of the Gauss-Markov source under squared error distortion. We obtain the Gaussian approximation for the Gauss-Markov source with excess-distortion criterion for any distortion d>0, and we show that the dispersion has a reverse waterfilling representation. This is the first finite blocklength result for lossy compression of sources with memory. We prove that the finite blocklength rate-distortion function R(n,d,ϵ) approaches the rate-distortion function R(d) as R(n,d,ϵ) = R(d) + √(V(d)/n)Q^-1(ϵ) + o(1/√(n)), where V(d) is the dispersion, ϵ∈ (0,1) is the excess-distortion probability, and Q^-1 is the inverse of the Q-function. We give a reverse waterfilling integral representation for the dispersion V(d), which parallels that of the rate-distortion functions for Gaussian processes. Remarkably, for all 0 < d≤σ^2/(1+|a|)^2, R(n,d,ϵ) of the Gauss-Markov source coincides with that of Z_k, the i.i.d. Gaussian noise driving the process, up to the second-order term. Among novel technical tools developed in this paper is a sharp approximation of the eigenvalues of the covariance matrix of n samples of the Gauss-Markov source, and a construction of a typical set using the maximum likelihood estimate of the parameter a based on n observations.

READ FULL TEXT
research
12/07/2012

Lossy Compression via Sparse Linear Regression: Computationally Efficient Encoding and Decoding

We propose computationally efficient encoders and decoders for lossy com...
research
06/30/2019

From Parameter Estimation to Dispersion of Nonstationary Gauss-Markov Processes

This paper provides a precise error analysis for the maximum likelihood ...
research
03/15/2018

Reconstructing Gaussian sources by spatial sampling

Consider a Gaussian memoryless multiple source with m components with jo...
research
11/08/2022

Asymptotically Optimal Stochastic Lossy Coding of Markov Sources

An effective 'on-the-fly' mechanism for stochastic lossy coding of Marko...
research
02/03/2012

Lossy Compression via Sparse Linear Regression: Performance under Minimum-distance Encoding

We study a new class of codes for lossy compression with the squared-err...
research
12/09/2014

Circumventing the Curse of Dimensionality in Prediction: Causal Rate-Distortion for Infinite-Order Markov Processes

Predictive rate-distortion analysis suffers from the curse of dimensiona...
research
11/23/2021

Towards Empirical Sandwich Bounds on the Rate-Distortion Function

Rate-distortion (R-D) function, a key quantity in information theory, ch...

Please sign up or login with your details

Forgot password? Click here to reset