Variable Selection with Second-Generation P-Values

12/14/2020
by   Yi Zuo, et al.
0

Many statistical methods have been proposed for variable selection in the past century, but few perform this task well. The current standard bearers for variable selection include smoothly clipped absolute deviation (SCAD), adaptive lasso (AL), and minimax concave penalty with penalized linear unbiased selection (MC+). In practice, however, these algorithms often struggle to balance support recovery and parameter estimation, despite well-established oracle behavior for variable selection in certain settings. Here we report on a novel application of second-generation p-values (SGPVs) for variable selection, which we call Penalized regression with SGPVs (ProSGPV). This approach has tangible advantages in balancing support recovery and parameter estimation. The ProSGPV approach captures the true model at the best rate achieved by the current standards, is easier to implement in practice, and yields parameter estimates with the smallest mean absolute error. Even with strong collinearity in the feature space, the ProSGPV approach can maintain its good performance by using a simple pre-screening step. Here we report on extensive simulations and two real-world applications comparing these approaches. Our ProSGPV algorithm is a fast and intuitive approach for variable selection that leverages the advantages of second-generation p-values.

READ FULL TEXT

page 41

page 42

research
09/20/2021

Variable Selection in GLM and Cox Models with Second-Generation P-Values

Variable selection has become a pivotal choice in data analyses that imp...
research
07/17/2020

An Easy-to-Implement Hierarchical Standardization for Variable Selection Under Strong Heredity Constraint

For many practical problems, the regression models follow the strong her...
research
02/08/2016

DECOrrelated feature space partitioning for distributed sparse regression

Fitting statistical models is computationally challenging when the sampl...
research
05/25/2019

Selective inference after variable selection via multiscale bootstrap

A general resampling approach is considered for selective inference prob...
research
11/22/2018

Penalized least squares approximation methods and their applications to stochastic processes

We construct an objective function that consists of a quadratic approxim...
research
04/17/2020

A regularization approach for stable estimation of loss development factors

In this article, we show that a new penalty function, which we call log-...
research
07/29/2021

Sparse estimation for generalized exponential marked Hawkes process

We have established a sparse estimation method for the generalized expon...

Please sign up or login with your details

Forgot password? Click here to reset