Sparse Optimization on Measures with Over-parameterized Gradient Descent

07/24/2019
by   Lenaïc Chizat, et al.
0

Minimizing a convex function of a measure with a sparsity-inducing penalty is a typical problem arising, e.g., in sparse spikes deconvolution or two-layer neural networks training. We show that this problem can be solved by discretizing the measure and running non-convex gradient descent on the positions and weights of the particles. For measures on a d-dimensional manifold and under some non-degeneracy assumptions, this leads to a global optimization algorithm with a complexity scaling as (1/ϵ) in the desired accuracy ϵ, instead of ϵ^-d for convex methods. The key theoretical tools are a local convergence analysis in Wasserstein space and an analysis of a perturbed mirror descent in the space of measures. Our bounds involve quantities that are exponential in d which is unavoidable under our assumptions.

READ FULL TEXT
research
05/24/2018

On the Global Convergence of Gradient Descent for Over-parameterized Models using Optimal Transport

Many tasks in machine learning and signal processing can be solved by mi...
research
02/09/2023

Efficient displacement convex optimization with particle gradient descent

Particle gradient descent, which uses particles to represent a probabili...
research
06/11/2021

LocoProp: Enhancing BackProp via Local Loss Optimization

We study a local loss construction approach for optimizing neural networ...
research
10/05/2018

Local Stability and Performance of Simple Gradient Penalty mu-Wasserstein GAN

Wasserstein GAN(WGAN) is a model that minimizes the Wasserstein distance...
research
01/05/2019

Analysis of a Two-Layer Neural Network via Displacement Convexity

Fitting a function by using linear combinations of a large number N of `...
research
09/16/2016

Gradient Descent Learns Linear Dynamical Systems

We prove that gradient descent efficiently converges to the global optim...
research
02/10/2023

Approximation and Structured Prediction with Sparse Wasserstein Barycenters

We develop a general theoretical and algorithmic framework for sparse ap...

Please sign up or login with your details

Forgot password? Click here to reset