Robust empirical risk minimization via Newton's method

01/30/2023
by   Eirini Ioannou, et al.
0

We study a variant of Newton's method for empirical risk minimization, where at each iteration of the optimization algorithm, we replace the gradient and Hessian of the objective function by robust estimators taken from existing literature on robust mean estimation for multivariate data. After proving a general theorem about the convergence of successive iterates to a small ball around the population-level minimizer, we study consequences of our theory in generalized linear models, when data are generated from Huber's epsilon-contamination model and/or heavy-tailed distributions. We also propose an algorithm for obtaining robust Newton directions based on the conjugate gradient method, which may be more appropriate for high-dimensional settings, and provide conjectures about the convergence of the resulting algorithm. Compared to the robust gradient descent algorithm proposed by Prasad et al. (2020), our algorithm enjoys the faster rates of convergence for successive iterates often achieved by second-order algorithms for convex problems, i.e., quadratic convergence in a neighborhood of the optimum, with a stepsize that may be chosen adaptively via backtracking linesearch.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/19/2018

Robust Estimation via Robust Gradient Estimation

We provide a new computationally-efficient class of estimators for risk ...
research
03/04/2018

A Distributed Quasi-Newton Algorithm for Empirical Risk Minimization with Nonsmooth Regularization

In this paper, we propose a communication- and computation- efficient di...
research
12/12/2019

A Distributed Quasi-Newton Algorithm for Primal and Dual Regularized Empirical Risk Minimization

We propose a communication- and computation-efficient distributed optimi...
research
02/14/2021

Distributed Second Order Methods with Fast Rates and Compressed Communication

We develop several new communication-efficient second-order methods for ...
research
01/15/2022

Quasi-Newton acceleration of EM and MM algorithms via Broyden's method

The principle of majorization-minimization (MM) provides a general frame...
research
06/25/2021

Hessian informed mirror descent

Inspired by the recent paper (L. Ying, Mirror descent algorithms for min...
research
11/01/2021

Free Probability, Newton lilypads and Jacobians of neural networks

Gradient descent during the learning process of a neural network can be ...

Please sign up or login with your details

Forgot password? Click here to reset