Sparse Regression with Multi-type Regularized Feature Modeling

10/07/2018
by   Sander Devriendt, et al.
0

Within the statistical and machine learning literature, regularization techniques are often used to construct sparse (predictive) models. Most regularization strategies only work for data where all predictors are of the same type, such as Lasso regression for continuous predictors. However, many predictive problems involve different predictor types. We propose a multi-type Lasso penalty that acts on the objective function as a sum of subpenalties, one for each predictor type. As such, we perform predictor selection and level fusion within a predictor in a data-driven way, simultaneous with the parameter estimation process. We develop a new estimation strategy for convex predictive models with this multi-type penalty. Using the theory of proximal operators, our estimation procedure is computationally efficient, partitioning the overall optimization problem into easier to solve subproblems, specific for each predictor type and its associated penalty. The proposed SMuRF algorithm improves on existing solvers in both accuracy and computational efficiency. This is demonstrated with an extensive simulation study and the analysis of a case-study on insurance pricing analytics.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/14/2019

Estimation and Feature Selection in Mixtures of Generalized Linear Experts Models

Mixtures-of-Experts (MoE) are conditional mixture models that have shown...
research
11/09/2010

Strong rules for discarding predictors in lasso-type problems

We consider rules for discarding predictors in lasso regression and rela...
research
07/10/2021

Cluster Regularization via a Hierarchical Feature Regression

Prediction tasks with high-dimensional nonorthogonal predictor sets pose...
research
03/18/2014

On the Sensitivity of the Lasso to the Number of Predictor Variables

The Lasso is a computationally efficient regression regularization proce...
research
10/11/2017

Adaptive multi-penalty regularization based on a generalized Lasso path

For many algorithms, parameter tuning remains a challenging and critical...
research
01/05/2021

Comparing methods addressing multi-collinearity when developing prediction models

Clinical prediction models are developed widely across medical disciplin...
research
04/10/2019

Joint Manifold Diffusion for Combining Predictions on Decoupled Observations

We present a new predictor combination algorithm that improves a given t...

Please sign up or login with your details

Forgot password? Click here to reset