HPO-B: A Large-Scale Reproducible Benchmark for Black-Box HPO based on OpenML

06/11/2021
by   Sebastian Pineda-Arango, et al.
0

Hyperparameter optimization (HPO) is a core problem for the machine learning community and remains largely unsolved due to the significant computational resources required to evaluate hyperparameter configurations. As a result, a series of recent related works have focused on the direction of transfer learning for quickly fine-tuning hyperparameters on a dataset. Unfortunately, the community does not have a common large-scale benchmark for comparing HPO algorithms. Instead, the de facto practice consists of empirical protocols on arbitrary small-scale meta-datasets that vary inconsistently across publications, making reproducibility a challenge. To resolve this major bottleneck and enable a fair and fast comparison of black-box HPO methods on a level playing field, we propose HPO-B, a new large-scale benchmark in the form of a collection of meta-datasets. Our benchmark is assembled and preprocessed from the OpenML repository and consists of 176 search spaces (algorithms) evaluated sparsely on 196 datasets with a total of 6.4 million hyperparameter evaluations. For ensuring reproducibility on our benchmark, we detail explicit experimental protocols, splits, and evaluation measures for comparing methods for both non-transfer, as well as, transfer learning HPO.

READ FULL TEXT

page 21

page 22

research
09/27/2019

Learning search spaces for Bayesian optimization: Another view of hyperparameter transfer learning

Bayesian optimization (BO) is a successful methodology to optimize black...
research
06/06/2023

Quick-Tune: Quickly Learning Which Pretrained Model to Finetune and How

With the ever-increasing number of pretrained models, machine learning p...
research
09/30/2019

A Copula approach for hyperparameter transfer learning

Bayesian optimization (BO) is a popular methodology to tune the hyperpar...
research
07/10/2023

SigOpt Mulch: An Intelligent System for AutoML of Gradient Boosted Trees

Gradient boosted trees (GBTs) are ubiquitous models used by researchers,...
research
12/15/2020

Amazon SageMaker Automatic Model Tuning: Scalable Black-box Optimization

Tuning complex machine learning systems is challenging. Machine learning...
research
06/10/2021

Meta-Learning for Symbolic Hyperparameter Defaults

Hyperparameter optimization in machine learning (ML) deals with the prob...
research
07/18/2017

On the State of the Art of Evaluation in Neural Language Models

Ongoing innovations in recurrent neural network architectures have provi...

Please sign up or login with your details

Forgot password? Click here to reset