Toward Theoretical Guidance for Two Common Questions in Practical Cross-Validation based Hyperparameter Selection

01/12/2023
by   Parikshit Ram, et al.
0

We show, to our knowledge, the first theoretical treatments of two common questions in cross-validation based hyperparameter selection: (1) After selecting the best hyperparameter using a held-out set, we train the final model using all of the training data – since this may or may not improve future generalization error, should one do this? (2) During optimization such as via SGD (stochastic gradient descent), we must set the optimization tolerance ρ – since it trades off predictive accuracy with computation cost, how should one set it? Toward these problems, we introduce the hold-in risk (the error due to not using the whole training data), and the model class mis-specification risk (the error due to having chosen the wrong model class) in a theoretical view which is simple, general, and suggests heuristics that can be used when faced with a dataset instance. In proof-of-concept studies in synthetic data where theoretical quantities can be controlled, we show that these heuristics can, respectively, (1) always perform at least as well as always performing retraining or never performing retraining, (2) either improve performance or reduce computational overhead by 2× with no loss in predictive performance.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/18/2018

Optimizing for Generalization in Machine Learning with Cross-Validation Gradients

Cross-validation is the workhorse of modern applied statistics and machi...
research
06/11/2023

Blocked Cross-Validation: A Precise and Efficient Method for Hyperparameter Tuning

Hyperparameter tuning plays a crucial role in optimizing the performance...
research
02/13/2019

Differential Description Length for Hyperparameter Selection in Machine Learning

This paper introduces a new method for model selection and more generall...
research
06/08/2021

Stability and Generalization of Bilevel Programming in Hyperparameter Optimization

Recently, the (gradient-based) bilevel programming framework is widely u...
research
01/29/2020

Asymptotics of Cross-Validation

Cross validation is a central tool in evaluating the performance of mach...
research
12/11/2002

Theoretical Analyses of Cross-Validation Error and Voting in Instance-Based Learning

This paper begins with a general theory of error in cross-validation tes...
research
08/16/2019

Selection of Exponential-Family Random Graph Models via Held-Out Predictive Evaluation (HOPE)

Statistical models for networks with complex dependencies pose particula...

Please sign up or login with your details

Forgot password? Click here to reset