Stochastic Approximation of Smooth and Strongly Convex Functions: Beyond the O(1/T) Convergence Rate

01/27/2019
by   Moshe Y. Vardi, et al.
12

Stochastic approximation (SA) is a classical approach for stochastic convex optimization. Previous studies have demonstrated that the convergence rate of SA can be improved by introducing either smoothness or strong convexity condition. In this paper, we make use of smoothness and strong convexity simultaneously to boost the convergence rate. Let λ be the modulus of strong convexity, κ be the condition number, F_* be the minimal risk, and α>1 be some small constant. First, we demonstrate that, in expectation, an O(1/[λ T^α] + κ F_*/T) risk bound is attainable when T = Ω(κ^α). Thus, when F_* is small, the convergence rate could be faster than O(1/[λ T]) and approaches O(1/[λ T^α]) in the ideal case. Second, to further benefit from small risk, we show that, in expectation, an O(1/2^T/κ+F_*) risk bound is achievable. Thus, the excess risk reduces exponentially until reaching O(F_*), and if F_*=0, we obtain a global linear convergence. Finally, we emphasize that our proof is constructive and each risk bound is equipped with an efficient stochastic algorithm attaining that bound.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/07/2014

Binary Excess Risk for Smooth Convex Surrogates

In statistical learning theory, convex surrogates of the 0-1 loss are hi...
research
02/21/2017

Stochastic Composite Least-Squares Regression with convergence rate O(1/n)

We consider the minimization of composite objective functions composed o...
research
04/28/2021

FastAdaBelief: Improving Convergence Rate for Belief-based Adaptive Optimizer by Strong Convexity

The AdaBelief algorithm demonstrates superior generalization ability to ...
research
05/29/2019

A unified variance-reduced accelerated gradient method for convex optimization

We propose a novel randomized incremental gradient algorithm, namely, VA...
research
01/09/2022

Smooth Nested Simulation: Bridging Cubic and Square Root Convergence Rates in High Dimensions

Nested simulation concerns estimating functionals of a conditional expec...
research
11/18/2015

On the Global Linear Convergence of Frank-Wolfe Optimization Variants

The Frank-Wolfe (FW) optimization algorithm has lately re-gained popular...
research
03/20/2022

Convergence rates of the stochastic alternating algorithm for bi-objective optimization

Stochastic alternating algorithms for bi-objective optimization are cons...

Please sign up or login with your details

Forgot password? Click here to reset