Quick-Tune: Quickly Learning Which Pretrained Model to Finetune and How

06/06/2023
by   Sebastian Pineda-Arango, et al.
0

With the ever-increasing number of pretrained models, machine learning practitioners are continuously faced with which pretrained model to use, and how to finetune it for a new dataset. In this paper, we propose a methodology that jointly searches for the optimal pretrained model and the hyperparameters for finetuning it. Our method transfers knowledge about the performance of many pretrained models with multiple hyperparameter configurations on a series of datasets. To this aim, we evaluated over 20k hyperparameter configurations for finetuning 24 pretrained image classification models on 87 datasets to generate a large-scale meta-dataset. We meta-learn a multi-fidelity performance predictor on the learning curves of this meta-dataset and use it for fast hyperparameter optimization on new datasets. We empirically demonstrate that our resulting approach can quickly select an accurate pretrained model for a new dataset together with its optimal hyperparameters.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/17/2017

Learning to Warm-Start Bayesian Hyperparameter Optimization

Hyperparameter optimization undergoes extensive evaluations of validatio...
research
06/28/2018

Automatic Exploration of Machine Learning Experiments on OpenML

Understanding the influence of hyperparameters on the performance of a m...
research
06/23/2022

On the Generalizability and Predictability of Recommender Systems

While other areas of machine learning have seen more and more automation...
research
06/11/2021

HPO-B: A Large-Scale Reproducible Benchmark for Black-Box HPO based on OpenML

Hyperparameter optimization (HPO) is a core problem for the machine lear...
research
02/27/2020

Using a thousand optimization tasks to learn hyperparameter search strategies

We present TaskSet, a dataset of tasks for use in training and evaluatin...
research
01/27/2022

Consolidated learning – a domain-specific model-free optimization strategy with examples for XGBoost and MIMIC-IV

For many machine learning models, a choice of hyperparameters is a cruci...
research
02/07/2021

Hyperparameter Optimization with Differentiable Metafeatures

Metafeatures, or dataset characteristics, have been shown to improve the...

Please sign up or login with your details

Forgot password? Click here to reset