On the Tunability of Optimizers in Deep Learning

10/25/2019
by   Prabhu Teja Sivaprasad, et al.
8

There is no consensus yet on the question whether adaptive gradient methods like Adam are easier to use than non-adaptive optimization methods like SGD. In this work, we fill in the important, yet ambiguous concept of `ease-of-use' by defining an optimizer's tunability: How easy is it to find good hyperparameter configurations using automatic random hyperparameter search? We propose a practical and universal quantitative measure for optimizer tunability that can form the basis for a fair optimizer benchmark. Evaluating a variety of optimizers on an extensive set of standard datasets and architectures, we find that Adam is the most tunable for the majority of problems, especially with a low budget for hyperparameter tuning.

READ FULL TEXT
research
07/29/2018

Is One Hyperparameter Optimizer Enough?

Hyperparameter tuning is the black art of automatically finding a good c...
research
01/28/2022

Adaptive Optimizer for Automated Hyperparameter Optimization Problem

The choices of hyperparameters have critical effects on the performance ...
research
07/03/2020

Descending through a Crowded Valley – Benchmarking Deep Learning Optimizers

Choosing the optimizer is among the most crucial decisions of deep learn...
research
11/09/2021

The Role of Adaptive Optimizers for Honest Private Hyperparameter Selection

Hyperparameter optimization is a ubiquitous challenge in machine learnin...
research
12/15/2020

Squirrel: A Switching Hyperparameter Optimizer

In this short note, we describe our submission to the NeurIPS 2020 BBO c...
research
09/27/2022

Efficient Non-Parametric Optimizer Search for Diverse Tasks

Efficient and automated design of optimizers plays a crucial role in ful...
research
06/30/2021

What can linear interpolation of neural network loss landscapes tell us?

Studying neural network loss landscapes provides insights into the natur...

Please sign up or login with your details

Forgot password? Click here to reset