Supplementary Material for CDC Submission: On Influence of Ill-conditioned Regression Matrix on Hyper-parameter Estimators for Kernel-based Regularization Methods

03/30/2020
by   Yue Ju, et al.
0

In this paper, we focus on the influences of the condition number of the regression matrix upon the comparison between two hyper-parameter estimation methods: the empirical Bayes (EB) and the Stein's unbiased estimator with respect to the mean square error (MSE) related to output prediction (SUREy). We firstly show that the greatest power of the condition number of the regression matrix of SUREy cost function convergence rate upper bound is always one larger than that of EB cost function convergence rate upper bound. Meanwhile, EB and SUREy hyper-parameter estimators are both proved to be asymptotically normally distributed under suitable conditions. In addition, one ridge regression case is further investigated to show that when the condition number of the regression matrix goes to infinity, the asymptotic variance of SUREy estimator tends to be larger than that of EB estimator.

READ FULL TEXT
research
03/30/2020

On Effects of Condition Number of Regression Matrix upon Hyper-parameter Estimators for Kernel-based Regularization Methods

In this paper, we focus on the influences of the condition number of the...
research
03/30/2020

Supplementary Material for CDC Submission No. 1461

In this paper, we focus on the influences of the condition number of the...
research
09/25/2022

Asymptotic Theory for Regularized System Identification Part I: Empirical Bayes Hyper-parameter Estimator

Regularized system identification is the major advance in system identif...
research
05/25/2019

Fast and Accurate Gaussian Kernel Ridge Regression Using Matrix Decompositions for Preconditioning

This paper presents a method for building a preconditioner for a kernel ...
research
01/17/2023

A Distribution Free Truncated Kernel Ridge Regression Estimator and Related Spectral Analyses

It is well known that kernel ridge regression (KRR) is a popular nonpara...
research
01/29/2020

An Upper Bound of the Bias of Nadaraya-Watson Kernel Regression under Lipschitz Assumptions

The Nadaraya-Watson kernel estimator is among the most popular nonparame...
research
05/29/2019

The cost-free nature of optimally tuning Tikhonov regularizers and other ordered smoothers

We consider the problem of selecting the best estimator among a family o...

Please sign up or login with your details

Forgot password? Click here to reset