Examples of pathological dynamics of the subgradient method for Lipschitz path-differentiable functions

07/22/2020
by   Rodolfo Rios-Zertuche, et al.
0

We show that the vanishing stepsize subgradient method – widely adopted for machine learning applications – can display rather messy behavior even in the presence of favorable assumptions. We establish that convergence of bounded subgradient sequences may fail even with a Whitney stratifiable objective function satisfying the Kurdyka-Lojasiewicz inequality. Moreover, when the objective function is path-differentiable we show that various properties all may fail to occur: criticality of the limit points, convergence of the sequence, convergence in values, codimension one of the accumulation set, equality of the accumulation and essential accumulation sets, connectedness of the essential accumulation set, spontaneous slowdown, oscillation compensation, and oscillation perpendicularity to the accumulation set.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/29/2020

Long term dynamics of the subgradient method for Lipschitz path differentiable functions

We consider the long-term dynamics of the vanishing stepsize subgradient...
research
08/31/2023

Frank-Wolfe algorithm for DC optimization problem

In the present paper, we formulate two versions of Frank–Wolfe algorithm...
research
07/20/2020

Sequential Quadratic Optimization for Nonlinear Equality Constrained Stochastic Optimization

Sequential quadratic optimization algorithms are proposed for solving sm...
research
10/04/2018

Convergence of the ADAM algorithm from a Dynamical System Viewpoint

Adam is a popular variant of the stochastic gradient descent for finding...
research
06/09/2017

Global Convergence of the (1+1) Evolution Strategy

We establish global convergence of the (1+1)-ES algorithm, i.e., converg...
research
07/18/2020

Improved Convergence Speed of Fully Symmetric Learning Rules for Principal Component Analysis

Fully symmetric learning rules for principal component analysis can be d...
research
09/23/2019

Conservative set valued fields, automatic differentiation, stochastic gradient method and deep learning

The Clarke subdifferential is not suited to tackle nonsmooth deep learni...

Please sign up or login with your details

Forgot password? Click here to reset