From safe screening rules to working sets for faster Lasso-type solvers

03/21/2017
by   Mathurin Massias, et al.
0

Convex sparsity-promoting regularizations are ubiquitous in modern statistical learning. By construction, they yield solutions with few non-zero coefficients, which correspond to saturated constraints in the dual optimization formulation. Working set (WS) strategies are generic optimization techniques that consist in solving simpler problems that only consider a subset of constraints, whose indices form the WS. Working set methods therefore involve two nested iterations: the outer loop corresponds to the definition of the WS and the inner loop calls a solver for the subproblems. For the Lasso estimator a WS is a set of features, while for a Group Lasso it refers to a set of groups. In practice, WS are generally small in this context so the associated feature Gram matrix can fit in memory. Here we show that the Gauss-Southwell rule (a greedy strategy for block coordinate descent techniques) leads to fast solvers in this case. Combined with a working set strategy based on an aggressive use of so-called Gap Safe screening rules, we propose a solver achieving state-of-the-art performance on sparse learning problems. Results are presented on Lasso and multi-task Lasso estimators.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/21/2018

Dual Extrapolation for Faster Lasso Solvers

Convex sparsity-inducing regularizations are ubiquitous in high-dimensio...
research
02/16/2019

Screening Rules for Lasso with Non-Convex Sparse Regularizers

Leveraging on the convexity of the Lasso problem , screening rules help ...
research
02/19/2016

GAP Safe Screening Rules for Sparse-Group-Lasso

In high dimensional settings, sparse structures are crucial for efficien...
research
05/13/2015

Mind the duality gap: safer rules for the Lasso

Screening rules allow to early discard irrelevant variables from the opt...
research
02/09/2020

Learning High Order Feature Interactions with Fine Control Kernels

We provide a methodology for learning sparse statistical models that use...
research
11/09/2010

Strong rules for discarding predictors in lasso-type problems

We consider rules for discarding predictors in lasso regression and rela...
research
07/12/2019

Dual Extrapolation for Sparse Generalized Linear Models

Generalized Linear Models (GLM) form a wide class of regression and clas...

Please sign up or login with your details

Forgot password? Click here to reset