Newton Method with Variable Selection by the Proximal Gradient Method

11/30/2022
by   Ryosuke Shimmura, et al.
0

In sparse estimation, in which the sum of the loss function and the regularization term is minimized, methods such as the proximal gradient method and the proximal Newton method are applied. The former is slow to converge to a solution, while the latter converges quickly but is inefficient for problems such as group lasso problems. In this paper, we examine how to efficiently find a solution by finding the convergence destination of the proximal gradient method. However, the case in which the Lipschitz constant of the derivative of the loss function is unknown has not been studied theoretically, and only the Newton method has been proposed for the case in which the Lipschitz constant is known. We show that the Newton method converges when the Lipschitz constant is unknown and extend the theory. Furthermore, we propose a new quasi-Newton method that avoids Hessian calculations and improves efficiency, and we prove that it converges quickly, providing a theoretical guarantee. Finally, numerical experiments show that the proposed method can significantly improve the efficiency.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/27/2019

A sparse semismooth Newton based proximal majorization-minimization algorithm for nonconvex square-root-loss regression problems

In this paper, we consider high-dimensional nonconvex square-root-loss r...
research
03/30/2020

Non-asymptotic Superlinear Convergence of Standard Quasi-Newton Methods

In this paper, we study the non-asymptotic superlinear convergence rate ...
research
07/01/2016

Randomized block proximal damped Newton method for composite self-concordant minimization

In this paper we consider the composite self-concordant (CSC) minimizati...
research
12/03/2021

Regularized Newton Method with Global O(1/k^2) Convergence

We present a Newton-type method that converges fast from any initializat...
research
06/27/2014

Proximal Quasi-Newton for Computationally Intensive L1-regularized M-estimators

We consider the class of optimization problems arising from computationa...
research
08/16/2023

Convergence of Two-Layer Regression with Nonlinear Units

Large language models (LLMs), such as ChatGPT and GPT4, have shown outst...
research
04/10/2023

Über die Anwendung des Tschebyschew-Verfahrens zum Ausbau des Weierstraß-Kerner-Verfahrens

We extend the Weierstrass-Kerner method by applying the Chebychev method...

Please sign up or login with your details

Forgot password? Click here to reset