Log In Sign Up

Gaussian Approximation of Quantization Error for Estimation from Compressed Data

by   Alon Kipnis, et al.

We consider the distributional connection between the lossy compressed representation of a high-dimensional signal X using a random spherical code and the observation of X under an additive white Gaussian noise (AWGN). We show that the Wasserstein distance between a bitrate-R compressed version of X and its observation under an AWGN-channel of signal-to-noise ratio 2^2R-1 is sub-linear in the problem dimension. We utilize this fact to connect the risk of an estimator based on an AWGN-corrupted version of X to the risk attained by the same estimator when fed with its bitrate-R quantized version. We demonstrate the usefulness of this connection by deriving various novel results for inference problems under compression constraints, including noisy source coding and limited-bitrate parameter estimation.


Sample Complexity Bounds for Learning High-dimensional Simplices in Noisy Regimes

In this paper, we propose a sample complexity bound for learning a simpl...

A spin glass model for reconstructing nonlinearly encrypted signals corrupted by noise

An encryption of a signal s∈R^N is a random mapping sy=(y_1,...,y_M)^T...

A Parameter Choice Rule for Tikhonov Regularization Based on Predictive Risk

In this work, we propose a new criterion for choosing the regularization...

Sparse Normal Means Estimation with Sublinear Communication

We consider the problem of sparse normal means estimation in a distribut...

Fast Performance Evaluation of Linear Block Codes over Memoryless Continuous Channels

There are rising scenarios in communication systems, where the noises ex...

Reconstruction of Sparse Signals under Gaussian Noise and Saturation

Most compressed sensing algorithms do not account for the effect of satu...

Distributional Transform Based Information Reconciliation

In this paper we present an information reconciliation protocol designed...