ℓ_1-regression with Heavy-tailed Distributions

05/02/2018
by   Moshe Y. Vardi, et al.
0

In this paper, we consider the problem of linear regression with heavy-tailed distributions. Different from previous studies that use the squared loss to measure the performance, we choose the absolute loss, which is more robust in the presence of large prediction errors. To address the challenge that both the input and output could be heavy-tailed, we propose a truncated minimization problem, and demonstrate that it enjoys an O(√(d/n)) excess risk, where d is the dimensionality and n is the number of samples. Compared with traditional work on ℓ_1-regression, the main advantage of our result is that we achieve a high-probability risk bound without exponential moment conditions on the input and output. Our theoretical guarantee is derived from a novel combination of the PAC-Bayesian analysis and the covering number.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/06/2021

On Empirical Risk Minimization with Dependent and Heavy-Tailed Data

In this work, we establish risk bounds for the Empirical Risk Minimizati...
research
10/24/2017

Taming the heavy-tailed features by shrinkage and clipping

In this paper, we consider the generalized linear models (GLM) with heav...
research
02/25/2021

Distribution-Free Robust Linear Regression

We study random design linear regression with no assumptions on the dist...
research
11/29/2022

Outlier-Robust Sparse Mean Estimation for Heavy-Tailed Distributions

We study the fundamental task of outlier-robust mean estimation for heav...
research
05/20/2019

PAC-Bayes under potentially heavy tails

We derive PAC-Bayesian learning guarantees for heavy-tailed losses, and ...
research
12/07/2017

Dimension-free PAC-Bayesian bounds for matrices, vectors, and linear least squares regression

This paper is focused on dimension-free PAC-Bayesian bounds, under weak ...
research
09/19/2020

Suboptimality of Constrained Least Squares and Improvements via Non-Linear Predictors

We study the problem of predicting as well as the best linear predictor ...

Please sign up or login with your details

Forgot password? Click here to reset