Sample Complexity of Stochastic Variance-Reduced Cubic Regularization for Nonconvex Optimization

02/20/2018
by   Zhe Wang, et al.
0

The popular cubic regularization (CR) method converges with first- and second-order optimality guarantee for nonconvex optimization, but encounters a high sample complexity issue for solving large-scale problems. Various sub-sampling variants of CR have been proposed to improve the sample complexity.In this paper, we propose a stochastic variance-reduced cubic-regularized (SVRC) Newton's method under both sampling with and without replacement schemes. We characterize the per-iteration sample complexity bounds which guarantee the same rate of convergence as that of CR for nonconvex optimization. Furthermore, our method achieves a total Hessian sample complexity of O(N^8/11ϵ^-3/2) and O(N^3/4ϵ^-3/2) respectively under sampling without and with replacement, which improve that of CR as well as other sub-sampling variant methods via the variance reduction scheme. Our result also suggests that sampling without replacement yields lower sample complexity than that of sampling with replacement. We further compare the practical performance of SVRC with other cubic regularization methods via experiments.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/29/2018

Sample Efficient Stochastic Variance-Reduced Cubic Regularization Method

We propose a sample efficient stochastic variance-reduced cubic regulari...
research
10/09/2018

Cubic Regularization with Momentum for Nonconvex Optimization

Momentum is a popular technique to accelerate the convergence in practic...
research
03/07/2023

Margin theory for the scenario-based approach to robust optimization in high dimension

This paper deals with the scenario approach to robust optimization. This...
research
08/29/2023

Second-order methods for quartically-regularised cubic polynomials, with applications to high-order tensor methods

There has been growing interest in high-order tensor methods for nonconv...
research
10/30/2022

Robust Data Valuation via Variance Reduced Data Shapley

Data valuation, especially quantifying data value in algorithmic predict...
research
10/12/2015

Toward a Better Understanding of Leaderboard

The leaderboard in machine learning competitions is a tool to show the p...
research
08/21/2023

A Homogenization Approach for Gradient-Dominated Stochastic Optimization

Gradient dominance property is a condition weaker than strong convexity,...

Please sign up or login with your details

Forgot password? Click here to reset