Robust adaptive variable selection in ultra-high dimensional regression models based on the density power divergence loss

04/11/2020
by   Abhik Ghosh, et al.
0

We consider the problem of simultaneous model selection and the estimation of regression coefficients in high-dimensional linear regression models of non-polynomial order, an extremely important problem of the recent era. The adaptive penalty functions are used in this regard to achieve the oracle model selection property along with easier computational burden. However, since the usual adaptive procedures (e.g., adaptive LASSO) based on the squared error loss function is extremely non-robust in the presence of data contamination which are a common problem with large scale data, e.g., noisy gene expression data. In this paper, we present a regularization procedure for the ultra-high dimensional data using a robust loss function based on the popular density power divergence (DPD) measure along with the adaptive LASSO penalty. We theoretically study the robustness and large-sample properties of the proposed adaptive robust estimator for a general class of error distribution; in particular, we show that the proposed adaptive DPD-LASSO estimator is highly robust, satisfies the oracle model selection property, and the corresponding estimators of the regression coefficients are consistent and asymptotically normal under easily verifiable set of assumptions. Illustrations are also provided for the most useful normal error density.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/31/2020

On regularization methods based on Rényi's pseudodistances for sparse high-dimensional linear regression models

Several regularization methods have been considered over the last decade...
research
04/18/2023

The Adaptive τ-Lasso: Its Robustness and Oracle Properties

This paper introduces a new regularized version of the robust τ-regressi...
research
10/21/2019

High-dimensional robust approximated M-estimators for mean regression with asymmetric data

Asymmetry along with heteroscedasticity or contamination often occurs wi...
research
10/23/2016

Inertial Regularization and Selection (IRS): Sequential Regression in High-Dimension and Sparsity

In this paper, we develop a new sequential regression modeling approach ...
research
05/14/2019

Fast and robust model selection based on ranks

We consider the problem of identifying important predictors in large dat...
research
04/17/2020

A regularization approach for stable estimation of loss development factors

In this article, we show that a new penalty function, which we call log-...
research
02/10/2022

Loss-guided Stability Selection

In modern data analysis, sparse model selection becomes inevitable once ...

Please sign up or login with your details

Forgot password? Click here to reset