A simple data-driven method to optimise the penalty strengths of penalised models and its application to non-parametric smoothing

06/08/2022
by   Jens Thomas, et al.
0

Information of interest can often only be extracted from data by model fitting. When the functional form of such a model can not be deduced from first principles, one has to make a choice between different possible models. A common approach in such cases is to minimise the information loss in the model by trying to reduce the number of fit variables (or the model flexibility, respectively) as much as possible while still yielding an acceptable fit to the data. Model selection via the Akaike Information Criterion (AIC) provides such an implementation of Occam's razor. We argue that the same principles can be applied to optimise the penalty-strength of a penalised maximum-likelihood model. However, while in typical applications AIC is used to choose from a finite, discrete set of maximum-likelihood models the penalty optimisation requires to select out of a continuum of candidate models and these models violate the maximum-likelihood condition. We derive a generalised information criterion AICp that encompasses this case. It naturally involves the concept of effective free parameters which is very flexible and can be applied to any model, be it linear or non-linear, parametric or non-parametric, and with or without constraint equations on the parameters. We show that the generalised AICp allows an optimisation of any penalty-strength without the need of separate Monte-Carlo simulations. As an example application, we discuss the optimisation of the smoothing in non-parametric models which has many applications in astrophysics, like in dynamical modeling, spectral fitting or gravitational lensing.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/28/2018

Jensen-Shannon Divergence as a Goodness-of-Fit Measure for Maximum Likelihood Estimation and Curve Fitting

The coefficient of determination, known as R^2, is commonly used as a go...
research
12/20/2013

Non-parametric Bayesian modeling of complex networks

Modeling structure in complex networks using Bayesian non-parametrics ma...
research
12/04/2020

Information Complexity Criterion for Model Selection in Robust Regression Using A New Robust Penalty Term

Model selection is basically a process of finding the best model from th...
research
02/27/2007

The Loss Rank Principle for Model Selection

We introduce a new principle for model selection in regression and class...
research
08/04/2021

Improved Non-parametric Penalized Maximum Likelihood Estimation for Arbitrarily Censored Survival Data

Non-parametric maximum likelihood estimation encompasses a group of clas...
research
11/12/2019

Robust Fitting for Generalized Additive Models for Location, Scale and Shape

The validity of estimation and smoothing parameter selection for the wid...
research
07/09/2021

Bayesian Error-in-Variables Models for the Identification of Power Networks

The increasing integration of intermittent renewable generation, especia...

Please sign up or login with your details

Forgot password? Click here to reset