Stochastic subgradient method converges on tame functions

04/20/2018
by   Damek Davis, et al.
0

This work considers the question: what convergence guarantees does the stochastic subgradient method have in the absence of smoothness and convexity? We prove that the stochastic subgradient method, on any semialgebraic locally Lipschitz function, produces limit points that are all first-order stationary. More generally, our result applies to any function with a Whitney stratifiable graph. In particular, this work endows the stochastic subgradient method with rigorous convergence guarantees for a wide class of problems arising in data science---including all popular deep learning architectures.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/17/2018

Stochastic model-based minimization of weakly convex functions

We consider an algorithm that successively samples and minimizes stochas...
research
12/07/2020

Continuum Limit of Lipschitz Learning on Graphs

Tackling semi-supervised learning problems with graph-based methods have...
research
01/11/2022

An apocalypse-free first-order low-rank optimization algorithm

We consider the problem of minimizing a differentiable function with loc...
research
02/19/2021

AI-SARAH: Adaptive and Implicit Stochastic Recursive Gradient Methods

We present an adaptive stochastic variance reduced method with an implic...
research
01/26/2023

A Fully First-Order Method for Stochastic Bilevel Optimization

We consider stochastic unconstrained bilevel optimization problems when ...
research
05/20/2017

Stabilizing Adversarial Nets With Prediction Methods

Adversarial neural networks solve many important problems in data scienc...
research
08/25/2022

An apocalypse-free first-order low-rank optimization algorithm with at most one rank reduction attempt per iteration

We consider the problem of minimizing a differentiable function with loc...

Please sign up or login with your details

Forgot password? Click here to reset