Stable recovery and the coordinate small-ball behaviour of random vectors

04/17/2019
by   Shahar Mendelson, et al.
0

Recovery procedures in various application in Data Science are based on stable point separation. In its simplest form, stable point separation implies that if f is "far away" from 0, and one is given a random sample (f(Z_i))_i=1^m where a proportional number of the sample points may be corrupted by noise, that information is still enough to exhibit that f is far from 0. Stable point separation is well understood in the context of iid sampling, and to explore it for general sampling methods we introduce a new notion---the coordinate small-ball of a random vector X. Roughly put, this feature captures the number of "relatively large coordinates" of (|<TX,u_i>|)_i=1^m, where T:R^n →R^m is an arbitrary linear operator and (u_i)_i=1^m is any fixed orthonormal basis of R^m. We show that under the bare-minimum assumptions on X, and with high probability, many of the values |<TX,u_i>| are at least of the order T_S_2/√(m). As a result, the "coordinate structure" of TX exhibits the typical Euclidean norm of TX and does so in a stable way. One outcome of our analysis is that random sub-sampled convolutions satisfy stable point separation under minimal assumptions on the generating random vector---a fact that was known previously only in a highly restrictive setup, namely, for random vectors with iid subgaussian coordinates.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/16/2019

On the geometry of polytopes generated by heavy-tailed random vectors

We study the geometry of centrally-symmetric random polytopes, generated...
research
09/15/2022

Structure preservation via the Wasserstein distance

We show that under minimal assumptions on a random vector X∈ℝ^d, and wit...
research
09/04/2017

Extending the small-ball method

The small-ball method was introduced as a way of obtaining a high probab...
research
09/12/2013

Recovery guarantees for exemplar-based clustering

For a certain class of distributions, we prove that the linear programmi...
research
02/16/2022

The Pareto cover problem

We introduce the problem of finding a set B of k points in [0,1]^n such ...
research
09/23/2020

Convergence of Gibbs Sampling: Coordinate Hit-and-Run Mixes Fast

The Gibbs Sampler is a general method for sampling high-dimensional dist...
research
07/15/2023

Bulk Johnson-Lindenstrauss Lemmas

For a set X of N points in ℝ^D, the Johnson-Lindenstrauss lemma provides...

Please sign up or login with your details

Forgot password? Click here to reset