On the Power of Preconditioning in Sparse Linear Regression

06/17/2021
by   Jonathan Kelner, et al.
0

Sparse linear regression is a fundamental problem in high-dimensional statistics, but strikingly little is known about how to efficiently solve it without restrictive conditions on the design matrix. We consider the (correlated) random design setting, where the covariates are independently drawn from a multivariate Gaussian N(0,Σ) with Σ : n × n, and seek estimators ŵ minimizing (ŵ-w^*)^TΣ(ŵ-w^*), where w^* is the k-sparse ground truth. Information theoretically, one can achieve strong error bounds with O(k log n) samples for arbitrary Σ and w^*; however, no efficient algorithms are known to match these guarantees even with o(n) samples, without further assumptions on Σ or w^*. As far as hardness, computational lower bounds are only known with worst-case design matrices. Random-design instances are known which are hard for the Lasso, but these instances can generally be solved by Lasso after a simple change-of-basis (i.e. preconditioning). In this work, we give upper and lower bounds clarifying the power of preconditioning in sparse linear regression. First, we show that the preconditioned Lasso can solve a large class of sparse linear regression problems nearly optimally: it succeeds whenever the dependency structure of the covariates, in the sense of the Markov property, has low treewidth – even if Σ is highly ill-conditioned. Second, we construct (for the first time) random-design instances which are provably hard for an optimally preconditioned Lasso. In fact, we complete our treewidth classification by proving that for any treewidth-t graph, there exists a Gaussian Markov Random Field on this graph such that the preconditioned Lasso, with any choice of preconditioner, requires Ω(t^1/20) samples to recover O(log n)-sparse signals when covariates are drawn from this model.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/26/2023

Feature Adaptation for Sparse Linear Regression

Sparse linear regression is a central problem in high-dimensional statis...
research
03/05/2022

Distributional Hardness Against Preconditioned Lasso via Erasure-Robust Designs

Sparse linear regression with ill-conditioned Gaussian random designs is...
research
07/25/2017

Compressed Sparse Linear Regression

High-dimensional sparse linear regression is a basic problem in machine ...
research
10/29/2017

Distributional Consistency of Lasso by Perturbation Bootstrap

Least Absolute Shrinkage and Selection Operator or the Lasso, introduced...
research
04/24/2023

Estimation of sparse linear regression coefficients under L-subexponential covariates

We address a task of estimating sparse coefficients in linear regression...
research
01/24/2020

Imputation for High-Dimensional Linear Regression

We study high-dimensional regression with missing entries in the covaria...
research
06/16/2022

On the well-spread property and its relation to linear regression

We consider the robust linear regression model y = Xβ^* + η, where an ad...

Please sign up or login with your details

Forgot password? Click here to reset