The cost-free nature of optimally tuning Tikhonov regularizers and other ordered smoothers

05/29/2019
by   Pierre C. Bellec, et al.
0

We consider the problem of selecting the best estimator among a family of Tikhonov regularized estimators, or, alternatively, to select a linear combination of these regularizers that is as good as the best regularizer in the family. Our theory reveals that if the Tikhonov regularizers share the same penalty matrix with different tuning parameters, a convex procedure based on Q-aggregation achieves the mean square error of the best estimator, up to a small error term no larger than Cσ^2, where σ^2 is the noise level and C>0 is an absolute constant. Remarkably, the error term does not depend on the penalty matrix or the number of estimators as long as they share the same penalty matrix, i.e., it applies to any grid of tuning parameters, no matter how large the cardinality of the grid is. This reveals the surprising "cost-free" nature of optimally tuning Tikhonov regularizers, in striking contrast with the existing literature on aggregation of estimators where one typically has to pay a cost of σ^2(M) where M is the number of estimators in the family. The result holds, more generally, for any family of ordered linear smoothers. This encompasses Ridge regression as well as Principal Component Regression. The result is extended to the problem of tuning Tikhonov regularizers with different penalty matrices.

READ FULL TEXT
research
07/11/2021

Derivatives and residual distribution of regularized M-estimators with application to adaptive tuning

This paper studies M-estimators with gradient-Lipschitz loss function re...
research
08/09/2019

The general Nature of Saturated Designs

In a full two-level factorial experiment the design matrix is a Hadamard...
research
12/09/2017

Stochastic Restricted Biased Estimators in misspecified regression model with incomplete prior information

In this article, the analysis of misspecification was extended to the re...
research
08/26/2021

Comparing Classes of Estimators: When does Gradient Descent Beat Ridge Regression in Linear Models?

Modern methods for learning from data depend on many tuning parameters, ...
research
03/30/2020

On Effects of Condition Number of Regression Matrix upon Hyper-parameter Estimators for Kernel-based Regularization Methods

In this paper, we focus on the influences of the condition number of the...

Please sign up or login with your details

Forgot password? Click here to reset