Automatic differentiation of nonsmooth iterative algorithms

05/31/2022
by   Jérôme Bolte, et al.
0

Differentiation along algorithms, i.e., piggyback propagation of derivatives, is now routinely used to differentiate iterative solvers in differentiable programming. Asymptotics is well understood for many smooth problems but the nondifferentiable case is hardly considered. Is there a limiting object for nonsmooth piggyback automatic differentiation (AD)? Does it have any variational meaning and can it be used effectively in machine learning? Is there a connection with classical derivative? All these questions are addressed under appropriate nonexpansivity conditions in the framework of conservative derivatives which has proved useful in understanding nonsmooth AD. For nonsmooth piggyback iterations, we characterize the attractor set of nonsmooth piggyback iterations as a set-valued fixed point which remains in the conservative framework. This has various consequences and in particular almost everywhere convergence of classical derivatives. Our results are illustrated on parametric convex optimization problems with forward-backward, Douglas-Rachford and Alternating Direction of Multiplier algorithms as well as the Heavy-Ball method.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/12/2023

Understanding Automatic Differentiation Pitfalls

Automatic differentiation, also known as backpropagation, AD, autodiff, ...
research
08/05/2022

Fixed-Point Automatic Differentiation of Forward–Backward Splitting Algorithms for Partly Smooth Functions

A large class of non-smooth practical optimization problems can be writt...
research
01/26/2023

Automatic differentiation as an effective tool in Electrical Impedance Tomography

Determining physical properties inside an object without access to direc...
research
01/03/2021

The structure of conservative gradient fields

The classical Clarke subdifferential alone is inadequate for understandi...
research
09/23/2019

Conservative set valued fields, automatic differentiation, stochastic gradient method and deep learning

The Clarke subdifferential is not suited to tackle nonsmooth deep learni...
research
04/02/2018

The simple essence of automatic differentiation (Differentiable functional programming made easy)

Automatic differentiation (AD) in reverse mode (RAD) is a central compon...
research
04/02/2018

The simple essence of automatic differentiation

Automatic differentiation (AD) in reverse mode (RAD) is a central compon...

Please sign up or login with your details

Forgot password? Click here to reset