DEEP-BO for Hyperparameter Optimization of Deep Networks

05/23/2019
by   Hyunghun Cho, et al.
0

The performance of deep neural networks (DNN) is very sensitive to the particular choice of hyper-parameters. To make it worse, the shape of the learning curve can be significantly affected when a technique like batchnorm is used. As a result, hyperparameter optimization of deep networks can be much more challenging than traditional machine learning models. In this work, we start from well known Bayesian Optimization solutions and provide enhancement strategies specifically designed for hyperparameter optimization of deep networks. The resulting algorithm is named as DEEP-BO (Diversified, Early-termination-Enabled, and Parallel Bayesian Optimization). When evaluated over six DNN benchmarks, DEEP-BO easily outperforms or shows comparable performance with some of the well-known solutions including GP-Hedge, Hyperband, BOHB, Median Stopping Rule, and Learning Curve Extrapolation. The code used is made publicly available at https://github.com/snu-adsl/DEEP-BO.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/08/2020

Sherpa: Robust Hyperparameter Optimization for Machine Learning

Sherpa is a hyperparameter optimization library for machine learning mod...
research
09/20/2021

SMAC3: A Versatile Bayesian Optimization Package for Hyperparameter Optimization

Algorithm parameters, in particular hyperparameters of machine learning ...
research
10/10/2019

Probabilistic Rollouts for Learning Curve Extrapolation Across Hyperparameter Settings

We propose probabilistic models that can extrapolate learning curves of ...
research
04/21/2023

Tree-structured Parzen estimator: Understanding its algorithm components and their roles for better empirical performance

Recent advances in many domains require more and more complicated experi...
research
09/20/2019

Bayesian Optimization for Iterative Learning

The success of deep (reinforcement) learning systems crucially depends o...
research
05/27/2023

PFNs4BO: In-Context Learning for Bayesian Optimization

In this paper, we use Prior-data Fitted Networks (PFNs) as a flexible su...
research
10/10/2022

PyHopper – Hyperparameter optimization

Hyperparameter tuning is a fundamental aspect of machine learning resear...

Please sign up or login with your details

Forgot password? Click here to reset