Loss minimization and parameter estimation with heavy tails

07/07/2013
by   Daniel Hsu, et al.
0

This work studies applications and generalizations of a simple estimation technique that provides exponential concentration under heavy-tailed distributions, assuming only bounded low-order moments. We show that the technique can be used for approximate minimization of smooth and strongly convex losses, and specifically for least squares linear regression. For instance, our d-dimensional estimator requires just Õ(d(1/δ)) random samples to obtain a constant factor approximation to the optimal least squares loss with probability 1-δ, without requiring the covariates or noise to be bounded or subgaussian. We provide further applications to sparse linear regression and low-rank covariance matrix estimation with similar allowances on the noise and covariate distributions. The core technique is a generalization of the median-of-means estimator to arbitrary metric spaces.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/07/2018

Correlation Estimation System Minimization Compared to Least Squares Minimization in Simple Linear Regression

A general method of minimization using correlation coefficients and orde...
research
09/06/2022

A spectral least-squares-type method for heavy-tailed corrupted regression with unknown covariance & heterogeneous noise

We revisit heavy-tailed corrupted least-squares linear regression assumi...
research
09/30/2022

A note on centering in subsample selection for linear regression

Centering is a commonly used technique in linear regression analysis. Wi...
research
06/15/2022

Robust and Sparse Estimation of Linear Regression Coefficients with Heavy-tailed Noises and Covariates

Robust and sparse estimation of linear regression coefficients is invest...
research
08/24/2022

Outlier Robust and Sparse Estimation of Linear Regression Coefficients

We consider outlier robust and sparse estimation of linear regression co...
research
08/31/2020

Estimating Rank-One Spikes from Heavy-Tailed Noise via Self-Avoiding Walks

We study symmetric spiked matrix models with respect to a general class ...
research
09/19/2020

Suboptimality of Constrained Least Squares and Improvements via Non-Linear Predictors

We study the problem of predicting as well as the best linear predictor ...

Please sign up or login with your details

Forgot password? Click here to reset