Proximal Quasi-Newton for Computationally Intensive L1-regularized M-estimators

06/27/2014
by   Kai Zhong, et al.
0

We consider the class of optimization problems arising from computationally intensive L1-regularized M-estimators, where the function or gradient values are very expensive to compute. A particular instance of interest is the L1-regularized MLE for learning Conditional Random Fields (CRFs), which are a popular class of statistical models for varied structured prediction problems such as sequence labeling, alignment, and classification with label taxonomy. L1-regularized MLEs for CRFs are particularly expensive to optimize since computing the gradient values requires an expensive inference step. In this work, we propose the use of a carefully constructed proximal quasi-Newton algorithm for such computationally intensive M-estimation problems, where we employ an aggressive active set selection technique. In a key contribution of the paper, we show that the proximal quasi-Newton method is provably super-linearly convergent, even in the absence of strong convexity, by leveraging a restricted variant of strong convexity. In our experiments, the proposed algorithm converges considerably faster than current state-of-the-art on the problems of sequence labeling and hierarchical classification.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/11/2016

Proximal Quasi-Newton Methods for Regularized Convex Optimization with Linear and Accelerated Sublinear Convergence Rates

In [19], a general, inexact, efficient proximal quasi-Newton algorithm f...
research
03/09/2023

Provably Convergent Plug-and-Play Quasi-Newton Methods

Plug-and-Play (PnP) methods are a class of efficient iterative methods t...
research
01/12/2020

The Proximal Method of Multipliers for a Class of Nonsmooth Convex Optimization

This paper develops the proximal method of multipliers for a class of no...
research
11/30/2022

Newton Method with Variable Selection by the Proximal Gradient Method

In sparse estimation, in which the sum of the loss function and the regu...
research
04/11/2017

Solving the L1 regularized least square problem via a box-constrained smooth minimization

In this paper, an equivalent smooth minimization for the L1 regularized ...
research
04/30/2020

One-Step Estimation With Scaled Proximal Methods

We study statistical estimators computed using iterative optimization me...
research
09/08/2015

A Scalable and Extensible Framework for Superposition-Structured Models

In many learning tasks, structural models usually lead to better interpr...

Please sign up or login with your details

Forgot password? Click here to reset