Necessary and Sufficient Conditions for Adaptive, Mirror, and Standard Gradient Methods

09/23/2019
by   Daniel Lévy, et al.
0

We study the impact of the constraint set and gradient geometry on the convergence of online and stochastic methods for convex optimization, providing a characterization of the geometries for which stochastic gradient and adaptive gradient methods are (minimax) optimal. In particular, we show that when the constraint set is quadratically convex, diagonally pre-conditioned stochastic gradient methods are minimax optimal. We further provide a converse that shows that when the constraints are not quadratically convex—for example, any ℓ_p-ball for p < 2—the methods are far from optimal. Based on this, we can provide concrete recommendations for when one should use adaptive, mirror or stochastic gradient methods.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/26/2017

Accelerating Stochastic Gradient Descent

There is widespread sentiment that it is not possible to effectively uti...
research
04/19/2019

Minimax Optimal Online Stochastic Learning for Sequences of Convex Functions under Sub-Gradient Observation Failures

We study online convex optimization under stochastic sub-gradient observ...
research
05/21/2018

On the Convergence of Stochastic Gradient Descent with Adaptive Stepsizes

Stochastic gradient descent is the method of choice for large scale opti...
research
05/23/2018

Adaptive Stochastic Gradient Langevin Dynamics: Taming Convergence and Saddle Point Escape Time

In this paper, we propose a new adaptive stochastic gradient Langevin dy...
research
11/06/2018

Double Adaptive Stochastic Gradient Optimization

Adaptive moment methods have been remarkably successful in deep learning...
research
08/04/2015

Asynchronous stochastic convex optimization

We show that asymptotically, completely asynchronous stochastic gradient...
research
07/26/2021

Convergence in quadratic mean of averaged stochastic gradient algorithms without strong convexity nor bounded gradient

Online averaged stochastic gradient algorithms are more and more studied...

Please sign up or login with your details

Forgot password? Click here to reset