A Robust Consistent Information Criterion for Model Selection based on Empirical Likelihood

by   Chixiang Chen, et al.

Conventional likelihood-based information criteria for model selection rely on the distribution assumption of data. However, for complex data that are increasingly available in many scientific fields, the specification of their underlying distribution turns out to be challenging, and the existing criteria may be limited and are not general enough to handle a variety of model selection problems. Here, we propose a robust and consistent model selection criterion based upon the empirical likelihood function which is data-driven. In particular, this framework adopts plug-in estimators that can be achieved by solving external estimating equations, not limited to the empirical likelihood, which avoids potential computational convergence issues and allows versatile applications, such as generalized linear models, generalized estimating equations, penalized regressions and so on. The formulation of our proposed criterion is initially derived from the asymptotic expansion of the marginal likelihood under variable selection framework, but more importantly, the consistent model selection property is established under a general context. Extensive simulation studies confirm the out-performance of the proposal compared to traditional model selection criteria. Finally, an application to the Atherosclerosis Risk in Communities Study illustrates the practical value of this proposed framework.



page 1

page 2

page 3

page 4


Empirical-likelihood-based criteria for model selection on marginal analysis of longitudinal data with dropout missingness

Longitudinal data are common in clinical trials and observational studie...

On model selection criteria for climate change impact studies

Climate change impact studies inform policymakers on the estimated damag...

General model-free weighted envelope estimation

Envelope methodology is succinctly pitched as a class of procedures for ...

Empirical bias-reducing adjustments to estimating functions

We develop a novel, general framework for the asymptotic reduction of th...

Identifying important predictors in large data bases – multiple testing and model selection

This is a chapter of the forthcoming Handbook of Multiple Testing. We co...

Towards a Theoretical Framework of Out-of-Distribution Generalization

Generalization to out-of-distribution (OOD) data, or domain generalizati...
This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.