Fast Regression for Structured Inputs

03/14/2022
โˆ™
by   Raphael A. Meyer, et al.
โˆ™
0
โˆ™

We study the โ„“_p regression problem, which requires finding ๐ฑโˆˆโ„^d that minimizes ๐€๐ฑ-๐›_p for a matrix ๐€โˆˆโ„^n ร— d and response vector ๐›โˆˆโ„^n. There has been recent interest in developing subsampling methods for this problem that can outperform standard techniques when n is very large. However, all known subsampling approaches have run time that depends exponentially on p, typically, d^๐’ช(p), which can be prohibitively expensive. We improve on this work by showing that for a large class of common structured matrices, such as combinations of low-rank matrices, sparse matrices, and Vandermonde matrices, there are subsampling based methods for โ„“_p regression that depend polynomially on p. For example, we give an algorithm for โ„“_p regression on Vandermonde matrices that runs in time ๐’ช(nlog^3 n+(dp^2)^0.5+ฯ‰ยทpolylog n), where ฯ‰ is the exponent of matrix multiplication. The polynomial dependence on p crucially allows our algorithms to extend naturally to efficient algorithms for โ„“_โˆž regression, via approximation of โ„“_โˆž by โ„“_๐’ช(log n). Of practical interest, we also develop a new subsampling algorithm for โ„“_p regression for arbitrary matrices, which is simpler than previous approaches for p โ‰ฅ 4.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset