On overfitting and post-selection uncertainty assessments

12/06/2017
by   Liang Hong, et al.
0

In a regression context, when the relevant subset of explanatory variables is uncertain, it is common to use a data-driven model selection procedure. Classical linear model theory, applied naively to the selected sub-model, may not be valid because it ignores the selected sub-model's dependence on the data. We provide an explanation of this phenomenon, in terms of overfitting, for a class of model selection criteria.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/25/2013

Exact post-selection inference, with application to the lasso

We develop a general approach to valid inference after model selection. ...
research
12/02/2020

The temporal overfitting problem with applications in wind power curve modeling

This paper is concerned with a nonparametric regression problem in which...
research
04/30/2015

Model Selection and Overfitting in Genetic Programming: Empirical Study [Extended Version]

Genetic Programming has been very successful in solving a large area of ...
research
11/18/2020

Post-Selection Inference via Algorithmic Stability

Modern approaches to data analysis make extensive use of data-driven mod...
research
01/27/2017

Subset Selection for Multiple Linear Regression via Optimization

Subset selection in multiple linear regression is to choose a subset of ...
research
10/31/2019

The consequences of checking for zero-inflation and overdispersion in the analysis of count data

Count data are ubiquitous in ecology and the Poisson generalized linear ...
research
01/19/2017

Parameter Selection Algorithm For Continuous Variables

In this article, we propose a new algorithm for supervised learning meth...

Please sign up or login with your details

Forgot password? Click here to reset