Smoothing the Edges: A General Framework for Smooth Optimization in Sparse Regularization using Hadamard Overparametrization

07/07/2023
by   Chris Kolb, et al.
0

This paper introduces a smooth method for (structured) sparsity in ℓ_q and ℓ_p,q regularized optimization problems. Optimization of these non-smooth and possibly non-convex problems typically relies on specialized procedures. In contrast, our general framework is compatible with prevalent first-order optimization methods like Stochastic Gradient Descent and accelerated variants without any required modifications. This is accomplished through a smooth optimization transfer, comprising an overparametrization of selected model parameters using Hadamard products and a change of penalties. In the overparametrized problem, smooth and convex ℓ_2 regularization of the surrogate parameters induces non-smooth and non-convex ℓ_q or ℓ_p,q regularization in the original parametrization. We show that our approach yields not only matching global minima but also equivalent local minima. This is particularly useful in non-convex sparse regularization, where finding global minima is NP-hard and local minima are known to generalize well. We provide a comprehensive overview consolidating various literature strands on sparsity-inducing parametrizations and propose meaningful extensions to existing approaches. The feasibility of our approach is evaluated through numerical experiments, which demonstrate that its performance is on par with or surpasses commonly used implementations of convex and non-convex regularization methods.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/24/2021

Why Do Local Methods Solve Nonconvex Problems?

Non-convex optimization is ubiquitous in modern machine learning. Resear...
research
03/21/2022

A Local Convergence Theory for the Stochastic Gradient Descent Method in Non-Convex Optimization With Non-isolated Local Minima

Non-convex loss functions arise frequently in modern machine learning, a...
research
08/31/2023

An Efficient Framework for Global Non-Convex Polynomial Optimization over the Hypercube

We present a novel efficient theoretical and numerical framework for sol...
research
08/03/2011

Optimization with Sparsity-Inducing Penalties

Sparse estimation methods are aimed at using or obtaining parsimonious r...
research
05/18/2021

Make $\ell_1$ Regularization Effective in Training Sparse CNN

Compressed Sensing using 𝓁1 regularization is among the most powerful an...
research
11/20/2017

Optimistic Robust Optimization With Applications To Machine Learning

Robust Optimization has traditionally taken a pessimistic, or worst-case...
research
11/23/2019

A Stochastic Tensor Method for Non-convex Optimization

We present a stochastic optimization method that uses a fourth-order reg...

Please sign up or login with your details

Forgot password? Click here to reset