Lossless Analog Compression

03/19/2018
by   Giovanni Alberti, et al.
0

We establish the fundamental limits of lossless analog compression by considering the recovery of arbitrary m-dimensional real random vectors x from the noiseless linear measurements y=Ax with n x m measurement matrix A. Our theory is inspired by the groundbreaking work of Wu and Verdu (2010) on almost lossless analog compression, but applies to the nonasymptotic, i.e., fixed-m case, and considers zero error probability. Specifically, our achievability result states that, for almost all A, the random vector x can be recovered with zero error probability provided that n > K(x), where the description complexity K(x) is given by the infimum of the lower modified Minkowski dimensions over all support sets U of x. We then particularize this achievability result to the class of s-rectifiable random vectors as introduced in Koliander et al. (2016); these are random vectors of absolutely continuous distribution---with respect to the s-dimensional Hausdorff measure---supported on countable unions of s-dimensional differentiable manifolds. Countable unions of differentiable manifolds include essentially all signal models used in compressed sensing theory, in spectrum-blind sampling, and in the matrix completion problem. Specifically, we prove that, for almost all A, s-rectifiable random vectors x can be recovered with zero error probability from n>s linear measurements. This threshold is, however, found not to be tight as exemplified by the construction of an s-rectifiable random vector that can be recovered with zero error probability from n<s linear measurements. This leads us to the introduction of the new class of s-analytic random vectors, which admit a strong converse in the sense of n greater than or equal to s being necessary for recovery with probability of error smaller than one. The central conceptual tool in the development of our theory is geometric measure theory.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/07/2023

Completion of Matrices with Low Description Complexity

We propose a theory for matrix completion that goes beyond the low-rank ...
research
10/30/2019

Superset Technique for Approximate Recovery in One-Bit Compressed Sensing

One-bit compressed sensing (1bCS) is a method of signal acquisition unde...
research
06/20/2019

Universality in Learning from Linear Measurements

We study the problem of recovering a structured signal from independentl...
research
07/18/2020

Compressed sensing of low-rank plus sparse matrices

Expressing a matrix as the sum of a low-rank matrix plus a sparse matrix...
research
12/19/2018

Derandomizing compressed sensing with combinatorial design

Compressed sensing is the art of reconstructing structured n-dimensional...
research
01/29/2019

Blind Unwrapping of Modulo Reduced Gaussian Vectors: Recovering MSBs from LSBs

We consider the problem of recovering n i.i.d samples from a zero mean m...
research
09/19/2018

Analog Coding Frame-work

Analog coding is a low-complexity method to combat erasures, based on li...

Please sign up or login with your details

Forgot password? Click here to reset