Exhuming nonnegative garrote from oblivion using suitable initial estimates- illustration in low and high-dimensional real data

10/27/2022
by   Edwin Kipruto, et al.
0

The nonnegative garrote (NNG) is among the first approaches that combine variable selection and shrinkage of regression estimates. When more than the derivation of a predictor is of interest, NNG has some conceptual advantages over the popular lasso. Nevertheless, NNG has received little attention. The original NNG relies on least-squares (OLS) estimates, which are highly variable in data with a high degree of multicollinearity (HDM) and do not exist in high-dimensional data (HDD). This might be the reason that NNG is not used in such data. Alternative initial estimates have been proposed but hardly used in practice. Analyzing three structurally different data sets, we demonstrated that NNG can also be applied in HDM and HDD and compared its performance with the lasso, adaptive lasso, relaxed lasso, and best subset selection in terms of variables selected, regression estimates, and prediction. Replacing OLS by ridge initial estimates in HDM and lasso initial estimates in HDD helped NNG select simpler models than competing approaches without much increase in prediction errors. Simpler models are easier to interpret, an important issue for descriptive modelling. Based on the limited experience from three datasets, we assume that the NNG can be a suitable alternative to the lasso and its extensions. Neutral comparison simulation studies are needed to better understand the properties of variable selection methods, compare them and derive guidance for practice.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/02/2019

Penalized Variable Selection in Multi-Parameter Regression Survival Modelling

Multi-parameter regression (MPR) modelling refers to the approach whereb...
research
07/30/2020

Solar: a least-angle regression for accurate and stable variable selection in high-dimensional data

We propose a new least-angle regression algorithm for variable selection...
research
12/21/2020

A critical review of LASSO and its derivatives for variable selection under dependence among covariates

We study the limitations of the well known LASSO regression as a variabl...
research
04/11/2019

FATSO: A family of operators for variable selection in linear models

In linear models it is common to have situations where several regressio...
research
03/28/2022

A Comparison of Hamming Errors of Representative Variable Selection Methods

Lasso is a celebrated method for variable selection in linear models, bu...
research
08/02/2018

High-dimensional regression in practice: an empirical study of finite-sample prediction, variable selection and ranking

Penalized likelihood methods are widely used for high-dimensional regres...
research
07/27/2017

Extended Comparisons of Best Subset Selection, Forward Stepwise Selection, and the Lasso

In exciting new work, Bertsimas et al. (2016) showed that the classical ...

Please sign up or login with your details

Forgot password? Click here to reset