Optimal Finite-Sum Smooth Non-Convex Optimization with SARAH

01/22/2019
by   Lam M. Nguyen, et al.
0

The total complexity (measured as the total number of gradient computations) of a stochastic first-order optimization algorithm that finds a first-order stationary point of a finite-sum smooth nonconvex objective function F(w)=1/n∑_i=1^n f_i(w) has been proven to be at least Ω(√(n)/ϵ) where ϵ denotes the attained accuracy E[ ∇ F(w̃)^2] ≤ϵ for the outputted approximation w̃ (Fang et al.,2018). This paper is the first to show that this lower bound is tight for the class of variance reduction methods which only assume the Lipschitz continuous gradient assumption. We prove this complexity result for a slightly modified version of the SARAH algorithm in (Nguyen et al.,2017a;b) - showing that SARAH is optimal and dominates all existing results. For convex optimization, we propose SARAH++ with sublinear convergence for general convex and linear convergence for strongly convex problems; and we provide a practical version for which numerical experiments on various datasets show an improved performance.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/22/2022

Simple and Optimal Stochastic Gradient Methods for Nonsmooth Nonconvex Optimization

We propose and analyze several stochastic gradient algorithms for findin...
research
03/21/2021

ANITA: An Optimal Loopless Accelerated Variance-Reduced Gradient Method

We propose a novel accelerated variance-reduced gradient method called A...
research
11/03/2022

Adaptive Stochastic Variance Reduction for Non-convex Finite-Sum Minimization

We propose an adaptive variance-reduction method, called AdaSpider, for ...
research
01/24/2021

An optimal gradient method for smooth (possibly strongly) convex minimization

We present an optimal gradient method for smooth (possibly strongly) con...
research
08/03/2023

Quantum speedups for stochastic optimization

We consider the problem of minimizing a continuous function given quantu...
research
11/07/2017

Convex Optimization with Nonconvex Oracles

In machine learning and optimization, one often wants to minimize a conv...
research
12/14/2021

Imaginary Zeroth-Order Optimization

Zeroth-order optimization methods are developed to overcome the practica...

Please sign up or login with your details

Forgot password? Click here to reset