Robust Regression Revisited: Acceleration and Improved Estimation Rates

06/22/2021
by   Arun Jambulapati, et al.
0

We study fast algorithms for statistical regression problems under the strong contamination model, where the goal is to approximately optimize a generalized linear model (GLM) given adversarially corrupted samples. Prior works in this line of research were based on the robust gradient descent framework of Prasad et. al., a first-order method using biased gradient queries, or the Sever framework of Diakonikolas et. al., an iterative outlier-removal method calling a stationary point finder. We present nearly-linear time algorithms for robust regression problems with improved runtime or estimation guarantees compared to the state-of-the-art. For the general case of smooth GLMs (e.g. logistic regression), we show that the robust gradient descent framework of Prasad et. al. can be accelerated, and show our algorithm extends to optimizing the Moreau envelopes of Lipschitz GLMs (e.g. support vector machines), answering several open questions in the literature. For the well-studied case of robust linear regression, we present an alternative approach obtaining improved estimation rates over prior nearly-linear time algorithms. Interestingly, our method starts with an identifiability proof introduced in the context of the sum-of-squares algorithm of Bakshi and Prasad, which achieved optimal error rates while requiring large polynomial runtime and sample complexity. We reinterpret their proof within the Sever framework and obtain a dramatically faster and more sample-efficient algorithm under fewer distributional assumptions.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/19/2019

PLS Generalized Linear Regression and Kernel Multilogit Algorithm (KMA) for Microarray Data Classification

We implement extensions of the partial least squares generalized linear ...
research
06/16/2021

Clustering Mixture Models in Almost-Linear Time via List-Decodable Mean Estimation

We study the problem of list-decodable mean estimation, where an adversa...
research
05/20/2023

A Novel Framework for Improving the Breakdown Point of Robust Regression Algorithms

We present an effective framework for improving the breakdown point of r...
research
10/01/2020

Agnostic Learning of Halfspaces with Gradient Descent via Soft Margins

We analyze the properties of gradient descent on convex surrogates for t...
research
01/30/2023

Near Optimal Private and Robust Linear Regression

We study the canonical statistical estimation problem of linear regressi...
research
11/30/2017

Outlier-robust moment-estimation via sum-of-squares

We develop efficient algorithms for estimating low-degree moments of unk...
research
06/18/2012

Linear Regression with Limited Observation

We consider the most common variants of linear regression, including Rid...

Please sign up or login with your details

Forgot password? Click here to reset