On Curvature-aided Incremental Aggregated Gradient Methods

05/31/2018 ∙ by Hoi-To Wai, et al. ∙ 0

This paper studies an acceleration technique for incremental aggregated gradient methods which exploits curvature information for solving strongly convex finite sum optimization problems. These optimization problems of interest arise in large-scale learning applications relevant to machine learning systems. The proposed methods utilizes a novel curvature-aided gradient tracking technique to produce gradient estimates using the aids of Hessian information during computation. We propose and analyze two curvature-aided methods --- the first method, called curvature-aided incremental aggregated gradient (CIAG) method, can be developed from the standard gradient method and it computes an ϵ-optimal solution using O( κ ( 1 / ϵ ) ) iterations for a small ϵ; the second method, called accelerated CIAG (A-CIAG) method, incorporates Nesterov's acceleration into CIAG and requires O( √(κ) ( 1 / ϵ ) ) iterations for a small ϵ, where κ is the problem's condition number. Importantly, the asymptotic convergence rates above are the same as those of the full gradient and accelerated full gradient methods, respectively, and they are independent of the number of component functions involved. The proposed methods are significantly faster than the state-of-the-art methods, especially for large-scale problems with a massive amount of data. The source codes are available at https://github.com/hoitowai/ciag/



There are no comments yet.


page 1

page 2

page 3

page 4

Code Repositories


companion codes for the paper "On Curvature-aided Incremental Aggregated Gradient Methods" submitted

view repo
This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.