Robust supervised learning with coordinate gradient descent

01/31/2022
by   Stéphane Gaïffas, et al.
0

This paper considers the problem of supervised learning with linear methods when both features and labels can be corrupted, either in the form of heavy tailed data and/or corrupted rows. We introduce a combination of coordinate gradient descent as a learning algorithm together with robust estimators of the partial derivatives. This leads to robust statistical learning methods that have a numerical complexity nearly identical to non-robust ones based on empirical risk minimization. The main idea is simple: while robust learning with gradient descent requires the computational cost of robustly estimating the whole gradient to update all parameters, a parameter can be updated immediately using a robust estimator of a single partial derivative in coordinate gradient descent. We prove upper bounds on the generalization error of the algorithms derived from this idea, that control both the optimization and statistical errors with and without a strong convexity assumption of the risk. Finally, we propose an efficient implementation of this approach in a new python library called linlearn, and demonstrate through extensive numerical experiments that our approach introduces a new interesting compromise between robustness, statistical performance and numerical efficiency for this problem.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/24/2023

Taylor Learning

Empirical risk minimization stands behind most optimization in supervise...
research
02/19/2018

Robust Estimation via Robust Gradient Estimation

We provide a new computationally-efficient class of estimators for risk ...
research
10/03/2019

Robust Risk Minimization for Statistical Learning

We consider a general statistical learning problem where an unknown frac...
research
09/07/2023

Empirical Risk Minimization for Losses without Variance

This paper considers an empirical risk minimization problem under heavy-...
research
05/20/2020

Model Repair: Robust Recovery of Over-Parameterized Statistical Models

A new type of robust estimation problem is introduced where the goal is ...
research
06/01/2017

Efficient learning with robust gradient descent

Minimizing the empirical risk is a popular training strategy, but for le...
research
08/10/2018

Ensemble Kalman Inversion: A Derivative-Free Technique For Machine Learning Tasks

The standard probabilistic perspective on machine learning gives rise to...

Please sign up or login with your details

Forgot password? Click here to reset