A first-order optimization algorithm for statistical learning with hierarchical sparsity structure

01/10/2020
by   Dewei Zhang, et al.
0

In many statistical learning problems, it is desired that the optimal solution conforms to an a priori known sparsity structure e.g. for better interpretability. Inducing such structures by means of convex regularizers requires nonsmooth penalty functions that exploit group overlapping. Our study focuses on evaluating the proximal operator of the Latent Overlapping Group lasso developed by Jacob et al. (2009). We develop an Alternating Direction Method of Multiplier with a sharing scheme to solve large-scale instance of the underlying optimization problem efficiently. In the absence of strong convexity, linear convergence of the algorithm is established using the error bound theory. More specifically, the paper contributes to establishing primal and dual error bounds over an unbounded feasible set and when the nonsmooth component in the objective function does not have a polyhedral epigraph. Numerical simulation studies supporting the proposed algorithm and two learning applications are discussed.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/07/2016

Linear Convergence of SVRG in Statistical Estimation

SVRG and its variants are among the state of art optimization algorithms...
research
06/26/2011

A General Framework for Structured Sparsity via Proximal Optimization

We study a generalized framework for structured sparsity. It extends the...
research
04/14/2022

Numerical evaluation of dual norms via the MM algorithm

We deal with the problem of numerically computing the dual norm, which i...
research
05/04/2011

Structured Sparsity via Alternating Direction Methods

We consider a class of sparse learning problems in high dimensional feat...
research
12/05/2015

Hierarchical Sparse Modeling: A Choice of Two Group Lasso Formulations

Demanding sparsity in estimated models has become a routine practice in ...
research
07/28/2018

Group-sparse SVD Models and Their Applications in Biological Data

Sparse Singular Value Decomposition (SVD) models have been proposed for ...
research
02/07/2021

Structured Sparsity Inducing Adaptive Optimizers for Deep Learning

The parameters of a neural network are naturally organized in groups, so...

Please sign up or login with your details

Forgot password? Click here to reset