Stochastic Quasi-Newton Methods for Nonconvex Stochastic Optimization

07/05/2016
by   Xiao Wang, et al.
0

In this paper we study stochastic quasi-Newton methods for nonconvex stochastic optimization, where we assume that noisy information about the gradients of the objective function is available via a stochastic first-order oracle (SFO). We propose a general framework for such methods, for which we prove almost sure convergence to stationary points and analyze its worst-case iteration complexity. When a randomly chosen iterate is returned as the output of such an algorithm, we prove that in the worst-case, the SFO-calls complexity is O(ϵ^-2) to ensure that the expectation of the squared norm of the gradient is smaller than the given accuracy tolerance ϵ. We also propose a specific algorithm, namely a stochastic damped L-BFGS (SdLBFGS) method, that falls under the proposed framework. Moreover, we incorporate the SVRG variance reduction technique into the proposed SdLBFGS method, and analyze its SFO-calls complexity. Numerical results on a nonconvex binary classification problem using SVM, and a multiclass classification problem using neural networks are reported.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/09/2018

A Stochastic Semismooth Newton Method for Nonsmooth Nonconvex Optimization

In this work, we present a globalized stochastic semismooth Newton metho...
research
03/07/2021

Retrospective Approximation for Smooth Stochastic Optimization

We consider stochastic optimization problems where a smooth (and potenti...
research
02/17/2020

Stochastic Gauss-Newton Algorithms for Nonconvex Compositional Optimization

We develop two new stochastic Gauss-Newton algorithms for solving a clas...
research
06/12/2021

Distributionally Robust Optimization with Markovian Data

We study a stochastic program where the probability distribution of the ...
research
01/16/2023

Faster Gradient-Free Algorithms for Nonsmooth Nonconvex Stochastic Optimization

We consider the optimization problem of the form min_x ∈ℝ^d f(x) ≜𝔼_ξ [F...
research
01/29/2020

Complexity Analysis of a Stochastic Cubic Regularisation Method under Inexact Gradient Evaluations and Dynamic Hessian Accuracy

We here adapt an extended version of the adaptive cubic regularisation m...
research
04/20/2023

Debiasing Conditional Stochastic Optimization

In this paper, we study the conditional stochastic optimization (CSO) pr...

Please sign up or login with your details

Forgot password? Click here to reset