CMA-ES for Post Hoc Ensembling in AutoML: A Great Success and Salvageable Failure

07/01/2023
by   Lennart Purucker, et al.
0

Many state-of-the-art automated machine learning (AutoML) systems use greedy ensemble selection (GES) by Caruana et al. (2004) to ensemble models found during model selection post hoc. Thereby, boosting predictive performance and likely following Auto-Sklearn 1's insight that alternatives, like stacking or gradient-free numerical optimization, overfit. Overfitting in Auto-Sklearn 1 is much more likely than in other AutoML systems because it uses only low-quality validation data for post hoc ensembling. Therefore, we were motivated to analyze whether Auto-Sklearn 1's insight holds true for systems with higher-quality validation data. Consequently, we compared the performance of covariance matrix adaptation evolution strategy (CMA-ES), state-of-the-art gradient-free numerical optimization, to GES on the 71 classification datasets from the AutoML benchmark for AutoGluon. We found that Auto-Sklearn's insight depends on the chosen metric. For the metric ROC AUC, CMA-ES overfits drastically and is outperformed by GES – statistically significantly for multi-class classification. For the metric balanced accuracy, CMA-ES does not overfit and outperforms GES significantly. Motivated by the successful application of CMA-ES for balanced accuracy, we explored methods to stop CMA-ES from overfitting for ROC AUC. We propose a method to normalize the weights produced by CMA-ES, inspired by GES, that avoids overfitting for CMA-ES and makes CMA-ES perform better than or similar to GES for ROC AUC.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/17/2023

Q(D)O-ES: Population-based Quality (Diversity) Optimisation for Post Hoc Ensemble Selection in AutoML

Automated machine learning (AutoML) systems commonly ensemble models pos...
research
02/16/2018

Train on Validation: Squeezing the Data Lemon

Model selection on validation data is an essential step in machine learn...
research
07/10/2018

Automatic Gradient Boosting

Automatic machine learning performs predictive modeling with high perfor...
research
02/10/2022

Heterogeneous Calibration: A post-hoc model-agnostic framework for improved generalization

We introduce the notion of heterogeneous calibration that applies a post...
research
05/25/2023

Interpretable Machine Learning based on Functional ANOVA Framework: Algorithms and Comparisons

In the early days of machine learning (ML), the emphasis was on developi...
research
06/15/2018

The Limits of Post-Selection Generalization

While statistics and machine learning offers numerous methods for ensuri...
research
06/06/2023

Bayesian post-hoc regularization of random forests

Random Forests are powerful ensemble learning algorithms widely used in ...

Please sign up or login with your details

Forgot password? Click here to reset