Near Optimal Heteroscedastic Regression with Symbiotic Learning

06/25/2023
โˆ™
by   Dheeraj Baby, et al.
โˆ™
0
โˆ™

We consider the problem of heteroscedastic linear regression, where, given n samples (๐ฑ_i, y_i) from y_i = โŸจ๐ฐ^*, ๐ฑ_i โŸฉ + ฯต_i ยทโŸจ๐Ÿ^*, ๐ฑ_i โŸฉ with ๐ฑ_i โˆผ N(0,๐ˆ), ฯต_i โˆผ N(0,1), we aim to estimate ๐ฐ^*. Beyond classical applications of such models in statistics, econometrics, time series analysis etc., it is also particularly relevant in machine learning when data is collected from multiple sources of varying but apriori unknown quality. Our work shows that we can estimate ๐ฐ^* in squared norm up to an error of ร•(๐Ÿ^*^2 ยท(1/n + (d/n)^2)) and prove a matching lower bound (upto log factors). This represents a substantial improvement upon the previous best known upper bound of ร•(๐Ÿ^*^2ยทd/n). Our algorithm is an alternating minimization procedure with two key subroutines 1. An adaptation of the classical weighted least squares heuristic to estimate ๐ฐ^*, for which we provide the first non-asymptotic guarantee. 2. A nonconvex pseudogradient descent procedure for estimating ๐Ÿ^* inspired by phase retrieval. As corollaries, we obtain fast non-asymptotic rates for two important problems, linear regression with multiplicative noise and phase retrieval with multiplicative noise, both of which are of independent interest. Beyond this, the proof of our lower bound, which involves a novel adaptation of LeCam's method for handling infinite mutual information quantities (thereby preventing a direct application of standard techniques like Fano's method), could also be of broader interest for establishing lower bounds for other heteroscedastic or heavy-tailed statistical problems.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
โˆ™ 05/31/2018

Efficient Algorithms and Lower Bounds for Robust Linear Regression

We study the problem of high-dimensional linear regression in a robust m...
research
โˆ™ 07/05/2021

Near-optimal inference in adaptive linear regression

When data is collected in an adaptive manner, even simple methods like o...
research
โˆ™ 09/06/2022

A spectral least-squares-type method for heavy-tailed corrupted regression with unknown covariance & heterogeneous noise

We revisit heavy-tailed corrupted least-squares linear regression assumi...
research
โˆ™ 10/31/2020

On Optimality of Meta-Learning in Fixed-Design Regression with Weighted Biased Regularization

We consider a fixed-design linear regression in the meta-learning model ...
research
โˆ™ 05/08/2023

Tight Bounds for Quantum Phase Estimation and Related Problems

Phase estimation, due to Kitaev [arXiv'95], is one of the most fundament...
research
โˆ™ 01/11/2021

Accurate error estimation in CG

In practical computations, the (preconditioned) conjugate gradient (P)CG...
research
โˆ™ 01/16/2017

High-Dimensional Regression with Binary Coefficients. Estimating Squared Error and a Phase Transition

We consider a sparse linear regression model Y=Xฮฒ^*+W where X has a Gaus...

Please sign up or login with your details

Forgot password? Click here to reset