Regularized Newton Method with Global O(1/k^2) Convergence

12/03/2021
by   Konstantin Mishchenko, et al.
0

We present a Newton-type method that converges fast from any initialization and for arbitrary convex objectives with Lipschitz Hessians. We achieve this by merging the ideas of cubic regularization with a certain adaptive Levenberg–Marquardt penalty. In particular, we show that the iterates given by x^k+1=x^k - (∇^2 f(x^k) + √(H∇ f(x^k))𝐈)^-1∇ f(x^k), where H>0 is a constant, converge globally with a 𝒪(1/k^2) rate. Our method is the first variant of Newton's method that has both cheap iterations and provably fast global convergence. Moreover, we prove that locally our method converges superlinearly when the objective is strongly convex. To boost the method's performance, we present a line search procedure that does not need hyperparameters and is provably efficient.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/01/2018

Global linear convergence of Newton's method without strong-convexity or Lipschitz gradients

We show that Newton's method converges globally at a linear rate for obj...
research
07/07/2020

A Distributed Cubic-Regularized Newton Method for Smooth Convex Optimization over Networks

We propose a distributed, cubic-regularized Newton method for large-scal...
research
08/11/2022

Super-Universal Regularized Newton Method

We analyze the performance of a variant of Newton method with quadratic ...
research
03/01/2020

A Globally Convergent Newton Method for Polynomials

Newton's method for polynomial root finding is one of mathematics' most ...
research
10/31/2022

An adaptive damped Newton method for strongly monotone and Lipschitz continuous operator equations

We will consider the damped Newton method for strongly monotone and Lips...
research
12/30/2013

Communication Efficient Distributed Optimization using an Approximate Newton-type Method

We present a novel Newton-type method for distributed optimization, whic...
research
11/30/2022

Newton Method with Variable Selection by the Proximal Gradient Method

In sparse estimation, in which the sum of the loss function and the regu...

Please sign up or login with your details

Forgot password? Click here to reset