On the well-spread property and its relation to linear regression

06/16/2022
by   Hongjie Chen, et al.
0

We consider the robust linear regression model y = Xβ^* + η, where an adversary oblivious to the design X ∈ℝ^n × d may choose η to corrupt all but a (possibly vanishing) fraction of the observations y in an arbitrary way. Recent work [dLN+21, dNS21] has introduced efficient algorithms for consistent recovery of the parameter vector. These algorithms crucially rely on the design matrix being well-spread (a matrix is well-spread if its column span is far from any sparse vector). In this paper, we show that there exists a family of design matrices lacking well-spreadness such that consistent recovery of the parameter vector in the above robust linear regression model is information-theoretically impossible. We further investigate the average-case time complexity of certifying well-spreadness of random matrices. We show that it is possible to efficiently certify whether a given n-by-d Gaussian matrix is well-spread if the number of observations is quadratic in the ambient dimension. We complement this result by showing rigorous evidence – in the form of a lower bound against low-degree polynomials – of the computational hardness of this same certification problem when the number of observations is o(d^2).

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/30/2020

Regress Consistently when Oblivious Outliers Overwhelm

We give a novel analysis of the Huber loss estimator for consistent robu...
research
06/17/2021

Statistical Query Lower Bounds for List-Decodable Linear Regression

We study the problem of list-decodable linear regression, where an adver...
research
11/06/2019

The gradient complexity of linear regression

We investigate the computational complexity of several basic linear alge...
research
08/04/2020

Well-Conditioned Methods for Ill-Conditioned Systems: Linear Regression with Semi-Random Noise

Classical iterative algorithms for linear system solving and regression ...
research
06/29/2022

Hardness and Algorithms for Robust and Sparse Optimization

We explore algorithms and limitations for sparse optimization problems s...
research
06/17/2021

On the Power of Preconditioning in Sparse Linear Regression

Sparse linear regression is a fundamental problem in high-dimensional st...
research
08/12/2022

Sparse change detection in high-dimensional linear regression

We introduce a new methodology 'charcoal' for estimating the location of...

Please sign up or login with your details

Forgot password? Click here to reset