Bag of Baselines for Multi-objective Joint Neural Architecture Search and Hyperparameter Optimization

05/03/2021
by   Julia Guerrero-Viu, et al.
0

Neural architecture search (NAS) and hyperparameter optimization (HPO) make deep learning accessible to non-experts by automatically finding the architecture of the deep neural network to use and tuning the hyperparameters of the used training pipeline. While both NAS and HPO have been studied extensively in recent years, NAS methods typically assume fixed hyperparameters and vice versa - there exists little work on joint NAS + HPO. Furthermore, NAS has recently often been framed as a multi-objective optimization problem, in order to take, e.g., resource requirements into account. In this paper, we propose a set of methods that extend current approaches to jointly optimize neural architectures and hyperparameters with respect to multiple objectives. We hope that these methods will serve as simple baselines for future research on multi-objective joint NAS + HPO. To facilitate this, all our code is available at https://github.com/automl/multi-obj-baselines.

READ FULL TEXT

page 12

page 22

07/30/2022

Tackling Neural Architecture Search With Quality Diversity Optimization

Neural architecture search (NAS) has been studied extensively and has gr...
07/18/2018

Towards Automated Deep Learning: Efficient Joint Neural Architecture and Hyperparameter Search

While existing work on neural architecture search (NAS) tunes hyperparam...
08/08/2022

Neural Architecture Search as Multiobjective Optimization Benchmarks: Problem Formulation and Performance Assessment

The ongoing advancements in network architecture design have led to rema...
06/10/2021

A multi-objective perspective on jointly tuning hardware and hyperparameters

In addition to the best model architecture and hyperparameters, a full A...
07/24/2020

What and Where: Learn to Plug Adapters via NAS for Multi-Domain Learning

As an important and challenging problem, multi-domain learning (MDL) typ...
05/27/2022

Auto-PINN: Understanding and Optimizing Physics-Informed Neural Architecture

Physics-informed neural networks (PINNs) are revolutionizing science and...
03/17/2022

Progressive Subsampling for Oversampled Data – Application to Quantitative MRI

We present PROSUB: PROgressive SUBsampling, a deep learning based, autom...

Code Repositories