Probabilistic Rollouts for Learning Curve Extrapolation Across Hyperparameter Settings

10/10/2019
by   Matilde Gargiani, et al.
0

We propose probabilistic models that can extrapolate learning curves of iterative machine learning algorithms, such as stochastic gradient descent for training deep networks, based on training data with variable-length learning curves. We study instantiations of this framework based on random forests and Bayesian recurrent neural networks. Our experiments show that these models yield better predictions than state-of-the-art models from the hyperparameter optimization literature when extrapolating the performance of neural networks trained with different hyperparameter settings.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/28/2020

Uncertainty Quantification in Deep Learning through Stochastic Maximum Principle

We develop a probabilistic machine learning method, which formulates a c...
research
05/23/2019

DEEP-BO for Hyperparameter Optimization of Deep Networks

The performance of deep neural networks (DNN) is very sensitive to the p...
research
01/06/2021

Hyperboost: Hyperparameter Optimization by Gradient Boosting surrogate models

Bayesian Optimization is a popular tool for tuning algorithms in automat...
research
03/06/2017

Forward and Reverse Gradient-Based Hyperparameter Optimization

We study two procedures (reverse-mode and forward-mode) for computing th...
research
10/21/2020

Learning Curves for Analysis of Deep Networks

A learning curve models a classifier's test error as a function of the n...
research
07/03/2019

Circuit-Based Intrinsic Methods to Detect Overfitting

The focus of this paper is on intrinsic methods to detect overfitting. T...
research
12/16/2017

NSML: A Machine Learning Platform That Enables You to Focus on Your Models

Machine learning libraries such as TensorFlow and PyTorch simplify model...

Please sign up or login with your details

Forgot password? Click here to reset