The out-of-sample R^2: estimation and inference

02/10/2023
by   Stijn Hawinkel, et al.
0

Out-of-sample prediction is the acid test of predictive models, yet an independent test dataset is often not available for assessment of the prediction error. For this reason, out-of-sample performance is commonly estimated using data splitting algorithms such as cross-validation or the bootstrap. For quantitative outcomes, the ratio of variance explained to total variance can be summarized by the coefficient of determination or in-sample R^2, which is easy to interpret and to compare across different outcome variables. As opposed to the in-sample R^2, the out-of-sample R^2 has not been well defined and the variability on the out-of-sample R̂^2 has been largely ignored. Usually only its point estimate is reported, hampering formal comparison of predictability of different outcome variables. Here we explicitly define the out-of-sample R^2 as a comparison of two predictive models, provide an unbiased estimator and exploit recent theoretical advances on uncertainty of data splitting estimates to provide a standard error for the R̂^2. The performance of the estimators for the R^2 and its standard error are investigated in a simulation study. We demonstrate our new method by constructing confidence intervals and comparing models for prediction of quantitative Brassica napus and Zea mays phenotypes based on gene expression data.

READ FULL TEXT
research
04/01/2021

Cross-validation: what does it estimate and how well does it do it?

Cross-validation is a widely-used technique to estimate prediction error...
research
01/09/2018

Test Error Estimation after Model Selection Using Validation Error

When performing supervised learning with the model selected using valida...
research
01/26/2022

Confidence intervals for the Cox model test error from cross-validation

Cross-validation (CV) is one of the most widely used techniques in stati...
research
04/25/2021

Model-based metrics: Sample-efficient estimates of predictive model subpopulation performance

Machine learning models - now commonly developed to screen, diagnose, or...
research
08/11/2019

Sample Splitting as an M-Estimator with Application to Physical Activity Scoring

Sample splitting is widely used in statistical applications, including c...
research
08/24/2020

Uncertainty in Bayesian Leave-One-Out Cross-Validation Based Model Comparison

Leave-one-out cross-validation (LOO-CV) is a popular method for comparin...
research
08/11/2019

Finite Sample Hypothesis Tests for Stacked Estimating Equations

Suppose there are two unknown parameters, each parameter is the solution...

Please sign up or login with your details

Forgot password? Click here to reset