Least-squares regressions via randomized Hessians

06/01/2020
by   Nabil Kahale, et al.
0

We consider the least-squares regression problem with a finite number of points. We analyze a novel approach, based on randomizing the Hessian matrix, to approximately solve this problem. The new algorithm is a variant of the averaged stochastic gradient descent method (SGD) with constant step-size. However, its updating rule relies on the entire response vector, and its convergence properties do not depend on the residuals. Without strong convexity assumptions, it is proven that the algorithm achieves a convergence rate for function values of O(1/k) after k iterations, where the constant behind the O notation does not depend explicitly on the smallest eigenvalue of the Hessian matrix. The algorithm has a preprocessing cost proportional to the input size, and the running time of each iteration is proportional to the dimension. In the strongly-convex case, a restart version of the algorithm yields a convergence rate of O(k^-l) in O(ld(n+k)) time for arbitrary l≥2, where the constant behind the O notation depends on l and on the smallest eigenvalue of the Hessian matrix. Our theoretical results are illustrated with numerical experiments.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/10/2013

Non-strongly-convex smooth stochastic approximation with convergence rate O(1/n)

We consider the stochastic approximation problem where a convex function...
research
03/02/2021

Convergence Rate of the (1+1)-Evolution Strategy with Success-Based Step-Size Adaptation on Convex Quadratic Functions

The (1+1)-evolution strategy (ES) with success-based step-size adaptatio...
research
05/11/2020

On Radial Isotropic Position: Theory and Algorithms

We review the theory of, and develop algorithms for transforming a finit...
research
10/16/2020

The conditioning of least squares problems in preconditioned variational data assimilation

Data assimilation algorithms combine prior and observational information...
research
02/21/2017

Stochastic Composite Least-Squares Regression with convergence rate O(1/n)

We consider the minimization of composite objective functions composed o...
research
01/13/2021

Learning with Gradient Descent and Weakly Convex Losses

We study the learning performance of gradient descent when the empirical...
research
10/25/2021

Maximizing the Smallest Eigenvalue of Grounded Laplacian Matrix

For a connected graph 𝒢=(V,E) with n nodes, m edges, and Laplacian matri...

Please sign up or login with your details

Forgot password? Click here to reset