Smooth Bilevel Programming for Sparse Regularization

06/02/2021
by   Clarice Poon, et al.
0

Iteratively reweighted least square (IRLS) is a popular approach to solve sparsity-enforcing regression problems in machine learning. State of the art approaches are more efficient but typically rely on specific coordinate pruning schemes. In this work, we show how a surprisingly simple reparametrization of IRLS, coupled with a bilevel resolution (instead of an alternating scheme) is able to achieve top performances on a wide range of sparsity (such as Lasso, group Lasso and trace norm regularizations), regularization strength (including hard constraints), and design matrices (ranging from correlated designs to differential operators). Similarly to IRLS, our method only involves linear systems resolutions, but in sharp contrast, corresponds to the minimization of a smooth function. Despite being non-convex, we show that there is no spurious minima and that saddle points are "ridable", so that there always exists a descent direction. We thus advocate for the use of a BFGS quasi-Newton solver, which makes our approach simple, robust and efficient. We perform a numerical benchmark of the convergence speed of our algorithm against state of the art solvers for Lasso, group Lasso, trace norm and linearly constrained problems. These results highlight the versatility of our approach, removing the need to use different solvers depending on the specificity of the ML problem under study.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/03/2022

Smooth over-parameterized solvers for non-smooth structured optimization

Non-smooth optimization is a core ingredient of many imaging or machine ...
research
11/27/2021

A dual semismooth Newton based augmented Lagrangian method for large-scale linearly constrained sparse group square-root Lasso problems

Square-root Lasso problems are proven robust regression problems. Furthe...
research
02/01/2019

On the Closed-form Proximal Mapping and Efficient Algorithms for Exclusive Lasso Models

The exclusive lasso regularization based on the ℓ_1,2 norm has become po...
research
06/08/2016

Efficient Smoothed Concomitant Lasso Estimation for High Dimensional Regression

In high dimensional settings, sparse structures are crucial for efficien...
research
10/24/2015

Fast and Scalable Lasso via Stochastic Frank-Wolfe Methods with a Convergence Guarantee

Frank-Wolfe (FW) algorithms have been often proposed over the last few y...
research
04/07/2011

Efficient First Order Methods for Linear Composite Regularizers

A wide class of regularization problems in machine learning and statisti...
research
11/02/2022

SIMD-size aware weight regularization for fast neural vocoding on CPU

This paper proposes weight regularization for a faster neural vocoder. P...

Please sign up or login with your details

Forgot password? Click here to reset