Stochastic subgradient method converges on tame functions

04/20/2018
by   Damek Davis, et al.
0

This work considers the question: what convergence guarantees does the stochastic subgradient method have in the absence of smoothness and convexity? We prove that the stochastic subgradient method, on any semialgebraic locally Lipschitz function, produces limit points that are all first-order stationary. More generally, our result applies to any function with a Whitney stratifiable graph. In particular, this work endows the stochastic subgradient method with rigorous convergence guarantees for a wide class of problems arising in data science---including all popular deep learning architectures.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset