Random Search Hyper-Parameter Tuning: Expected Improvement Estimation and the Corresponding Lower Bound

08/17/2022
by   Dan Navon, et al.
0

Hyperparameter tuning is a common technique for improving the performance of neural networks. Most techniques for hyperparameter search involve an iterated process where the model is retrained at every iteration. However, the expected accuracy improvement from every additional search iteration, is still unknown. Calculating the expected improvement can help create stopping rules for hyperparameter tuning and allow for a wiser allocation of a project's computational budget. In this paper, we establish an empirical estimate for the expected accuracy improvement from an additional iteration of hyperparameter search. Our results hold for any hyperparameter tuning method which is based on random search <cit.> and samples hyperparameters from a fixed distribution. We bound our estimate with an error of O(√(log k/k)) w.h.p. where k is the current number of iterations. To the best of our knowledge this is the first bound on the expected gain from an additional iteration of hyperparameter search. Finally, we demonstrate that the optimal estimate for the expected accuracy will still have an error of 1/k.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/21/2020

A Gradient-based Bilevel Optimization Approach for Tuning Hyperparameters in Machine Learning

Hyperparameter tuning is an active area of research in machine learning,...
research
06/06/2017

Random Search for Hyperparameters using Determinantal Point Processes

We propose the use of k-determinantal point processes in hyperparameter ...
research
12/23/2021

Using Sequential Statistical Tests to Improve the Performance of Random Search in hyperparameter Tuning

Hyperparamter tuning is one of the the most time-consuming parts in mach...
research
10/05/2020

Non-trivial informational closure of a Bayesian hyperparameter

We investigate the non-trivial informational closure (NTIC) of a Bayesia...
research
02/21/2019

Bayes Optimal Early Stopping Policies for Black-Box Optimization

We derive an optimal policy for adaptively restarting a randomized algor...
research
05/25/2019

Hyperparameter-Free Out-of-Distribution Detection Using Softmax of Scaled Cosine Similarity

The ability of detecting out-of-distribution (OOD) samples is important ...
research
09/16/2019

A Tsetlin Machine with Multigranular Clauses

The recently introduced Tsetlin Machine (TM) has provided competitive pa...

Please sign up or login with your details

Forgot password? Click here to reset