Online Robust Regression via SGD on the l1 loss

07/01/2020
by   Scott Pesme, et al.
0

We consider the robust linear regression problem in the online setting where we have access to the data in a streaming manner, one data point after the other. More specifically, for a true parameter θ^*, we consider the corrupted Gaussian linear model y = ⟨ x , θ^* ⟩ + ε + b where the adversarial noise b can take any value with probability η and equals zero otherwise. We consider this adversary to be oblivious (i.e., b independent of the data) since this is the only contamination model under which consistency is possible. Current algorithms rely on having the whole data at hand in order to identify and remove the outliers. In contrast, we show in this work that stochastic gradient descent on the ℓ_1 loss converges to the true parameter vector at a Õ( 1 / (1 - η)^2 n ) rate which is independent of the values of the contaminated measurements. Our proof relies on the elegant smoothing of the non-smooth ℓ_1 loss by the Gaussian data and a classical non-asymptotic analysis of Polyak-Ruppert averaged SGD. In addition, we provide experimental evidence of the efficiency of this simple and highly scalable algorithm.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/05/2018

Stochastic Gradient Descent on Separable Data: Exact Convergence with a Fixed Learning Rate

Stochastic Gradient Descent (SGD) is a central tool in machine learning....
research
12/22/2017

True Asymptotic Natural Gradient Optimization

We introduce a simple algorithm, True Asymptotic Natural Gradient Optimi...
research
09/29/2022

Statistical Learning and Inverse Problems: An Stochastic Gradient Approach

Inverse problems are paramount in Science and Engineering. In this paper...
research
01/10/2020

Choosing the Sample with Lowest Loss makes SGD Robust

The presence of outliers can potentially significantly skew the paramete...
research
06/06/2021

Fast and Robust Online Inference with Stochastic Gradient Descent via Random Scaling

We develop a new method of online inference for a vector of parameters e...
research
08/15/2023

Max-affine regression via first-order methods

We consider regression of a max-affine model that produces a piecewise l...
research
06/20/2021

Robust Regression via Model Based Methods

The mean squared error loss is widely used in many applications, includi...

Please sign up or login with your details

Forgot password? Click here to reset