Memory-Sample Tradeoffs for Linear Regression with Small Error

04/18/2019
by   Vatsal Sharan, et al.
0

We consider the problem of performing linear regression over a stream of d-dimensional examples, and show that any algorithm that uses a subquadratic amount of memory exhibits a slower rate of convergence than can be achieved without memory constraints. Specifically, consider a sequence of labeled examples (a_1,b_1), (a_2,b_2)..., with a_i drawn independently from a d-dimensional isotropic Gaussian, and where b_i = 〈 a_i, x〉 + η_i, for a fixed x ∈R^d with x_2 = 1 and with independent noise η_i drawn uniformly from the interval [-2^-d/5,2^-d/5]. We show that any algorithm with at most d^2/4 bits of memory requires at least Ω(d 1/ϵ) samples to approximate x to ℓ_2 error ϵ with probability of success at least 2/3, for ϵ sufficiently small as a function of d. In contrast, for such ϵ, x can be recovered to error ϵ with probability 1-o(1) with memory O(d^2 (1/ϵ)) using d examples. This represents the first nontrivial lower bounds for regression with super-linear memory, and may open the door for strong memory/sample tradeoffs for continuous optimization.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/31/2018

Efficient Algorithms and Lower Bounds for Robust Linear Regression

We study the problem of high-dimensional linear regression in a robust m...
research
06/17/2021

Statistical Query Lower Bounds for List-Decodable Linear Regression

We study the problem of list-decodable linear regression, where an adver...
research
09/20/2023

GLM Regression with Oblivious Corruptions

We demonstrate the first algorithms for the problem of regression for ge...
research
07/05/2021

Memory-Sample Lower Bounds for Learning Parity with Noise

In this work, we show, for the well-studied problem of learning parity u...
research
06/09/2022

Strong Memory Lower Bounds for Learning Natural Models

We give lower bounds on the amount of memory required by one-pass stream...
research
08/08/2017

Extractor-Based Time-Space Lower Bounds for Learning

A matrix M: A × X →{-1,1} corresponds to the following learning problem:...
research
03/31/2022

Learning from many trajectories

We initiate a study of supervised learning from many independent sequenc...

Please sign up or login with your details

Forgot password? Click here to reset