Moving Beyond Sub-Gaussianity in High-Dimensional Statistics: Applications in Covariance Estimation and Linear Regression

04/08/2018
by   Arun Kumar Kuchibhotla, et al.
0

Concentration inequalities form an essential toolkit in the study of high-dimensional statistical methods. Most of the relevant statistics literature is based on the assumptions of sub-Gaussian/sub-exponential random vectors. In this paper, we bring together various probability inequalities for sums of independent random variables under much weaker exponential type (sub-Weibull) tail assumptions. These results extract a part sub-Gaussian tail behavior in finite samples, matching the asymptotics governed by the central limit theorem, and are compactly represented in terms of a new Orlicz quasi-norm - the Generalized Bernstein-Orlicz norm - that typifies such tail behaviors. We illustrate the usefulness of these inequalities through the analysis of four fundamental problems in high-dimensional statistics. In the first two problems, we study the rate of convergence of the sample covariance matrix in terms of the maximum elementwise norm and the maximum k-sub-matrix operator norm which are key quantities of interest in bootstrap procedures and high-dimensional structured covariance matrix estimation. The third example concerns the restricted eigenvalue condition, required in high dimensional linear regression, which we verify for all sub-Weibull random vectors under only marginal (not joint) tail assumptions on the covariates. To our knowledge, this is the first unified result obtained in such generality. In the final example, we consider the Lasso estimator for linear regression and establish its rate of convergence under much weaker tail assumptions (on the errors as well as the covariates) than those in the existing literature. The common feature in all our results is that the convergence rates under most exponential tails match the usual ones under sub-Gaussian assumptions. Finally, we also establish a high-dimensional CLT and tail bounds for empirical processes for sub-Weibulls.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/23/2023

Concentration for high-dimensional linear processes with dependent innovations

We develop concentration inequalities for the l_∞ norm of a vector linea...
research
11/04/2020

Concentration Inequalities for Statistical Inference

This paper gives a review of concentration inequalities which are widely...
research
02/04/2021

Sharper Sub-Weibull Concentrations: Non-asymptotic Bai-Yin Theorem

Arising in high-dimensional probability, non-asymptotic concentration in...
research
02/07/2019

Tail behavior of dependent V-statistics and its applications

We establish exponential inequalities and Cramer-type moderate deviation...
research
02/12/2013

Covariance Estimation in High Dimensions via Kronecker Product Expansions

This paper presents a new method for estimating high dimensional covaria...
research
02/14/2023

Maximum interpoint distance of high-dimensional random vectors

A limit theorem for the largest interpoint distance of p independent and...
research
09/16/2020

Universality Laws for High-Dimensional Learning with Random Features

We prove a universality theorem for learning with random features. Our r...

Please sign up or login with your details

Forgot password? Click here to reset