GAMA: a General Automated Machine learning Assistant

by   Pieter Gijsbers, et al.
TU Eindhoven

The General Automated Machine learning Assistant (GAMA) is a modular AutoML system developed to empower users to track and control how AutoML algorithms search for optimal machine learning pipelines, and facilitate AutoML research itself. In contrast to current, often black-box systems, GAMA allows users to plug in different AutoML and post-processing techniques, logs and visualizes the search process, and supports easy benchmarking. It currently features three AutoML search algorithms, two model post-processing steps, and is designed to allow for more components to be added.


page 1

page 2

page 3

page 4


MimickNet, Matching Clinical Post-Processing Under Realistic Black-Box Constraints

Image post-processing is used in clinical-grade ultrasound scanners to i...

Multiaccuracy: Black-Box Post-Processing for Fairness in Classification

Machine learning predictors are successfully deployed in applications ra...

Physically constrained causal noise models for high-contrast imaging of exoplanets

The detection of exoplanets in high-contrast imaging (HCI) data hinges o...

Naive Automated Machine Learning

An essential task of Automated Machine Learning (AutoML) is the problem ...

Automated Machine Learning – a brief review at the end of the early years

Automated machine learning (AutoML) is the sub-field of machine learning...

Naive Automated Machine Learning – A Late Baseline for AutoML

Automated Machine Learning (AutoML) is the problem of automatically find...

1 Introduction

Automated Machine Learning (AutoML) aims to automate the process of building machine learning models, for instance, by automating the selection and tuning of preprocessing and learning algorithms in machine learning pipelines. In recent years, many AutoML systems have been developed, such as Auto-WEKA [10], auto-sklearn [4], TPOT [9] and ML-Plan [8]. They vary in the types of pipelines they build (e.g. fixed or variable length), how they optimize them (e.g. using evolutionary or Bayesian optimization), and whether or how they employ meta-learning (e.g. warm-starting) or post-processing (e.g. ensembling).

We demonstrate111A video demonstration can be found at

a new open-source AutoML system, GAMA

222Code and documentation can be found at [6], which distinguishes itself by it modularity (allowing users to compose AutoML systems from sub-components), extensibility (allowing new components to be added), transparency (tracking and visualizing the search process to better understand what the AutoML system is doing), and support for research, such as integration with the AutoML benchmark [5]. The main difference to our earlier publication ([6]) is the redesign to allow for a modular AutoML pipeline and the addition of a graphical user interface.

As such, it caters to a wide range of users, from people without a deep machine learning background who want an easy-to-use AutoML tool, to those who want better control and understanding of the AutoML process, and especially researchers who want to perform systematic AutoML research.

Currently, three different search algorithms and two post-processing techniques are available, but we welcome and plan to include more techniques in the future. For novice users, GAMA offers a default configuration shown to perform well in our benchmarks.

2 System Overview

Modular AutoML Pipeline

Rather than prescribing a specific combination of AutoML techniques, GAMA allows users to combine different search and post-processing algorithms into a flexible AutoML ‘pipeline’ that can be tuned to the problem at hand.

There are three optimization algorithms currently implemented in GAMA to search for optimal machine learning pipelines: random search [1], an asynchronous successive halving algorithm (ASHA) [7]

which uses low-fidelity estimates to filter out bad pipelines early, and an asynchronous multi-objective evolutionary algorithm.

After the pipeline search has completed, a post-processing technique will be executed to construct the final model. It is currently possible to either train the single best pipeline or create an ensemble out of pipelines evaluated during search, as described in [2]. In subsequent work, we plan to expand the number of search and post-processing techniques available out-of-the-box.

Listing 1 shows how to configure GAMA with non-default search and postprocessing methods and use it as a drop-in replacement for scikit-learn estimators.333An always up-to-date version of this listing can be found at

New AutoML algorithms or variations to existing ones can be included and tested with relative ease. For instance, each of the search algorithms described above has been implemented and integrated in GAMA with less than 170 lines of code, and they can all make use of shared functions for logging, parallel pipeline evaluation and adhering to runtime constraints. It also allows users to research other questions, such as how to choose the search algorithm for AutoML.

from gama import GamaClassifier
from gama.search_methods import AsynchronousSuccessiveHalving
from gama.postprocessing import EnsemblePostProcessing
automl = GamaClassifier(
), y)
automl.predict(X_test), y_test)
Listing 1: Configuring an AutoML pipeline with GAMA


GAMA comes with a graphical web interface which allows novice users to start and configure GAMA.Moreover, it visualizes the AutoML process to enable researchers to easily monitor and analyse the behavior of specific AutoML configurations.

GAMA logs the creation and evaluation of each pipeline, including meta-data such as creation time and evaluation duration. For pipelines created through evolution, it also records the parent pipelines and how they differ. One can also compare multiple logs at once, creating figures such as Figure 1 that shows the convergence rate of five different GAMA runs over time on the airline dataset444

Figure 1: Visualization of logs


GAMA in integrated with the open-source AutoML Benchmark introduced in [5]. Figure 2 shows the results of running GAMA with its default settings some of the biggest and most challenging datasets for which each other framework had results in the original work.555Although we could not run these experiments on the same (AWS) hardware, we took care to use the same computational constraints. The full and latest results will be made available in the GAMA documentation.

Figure 2: Performance benchmark results

3 Related work

GAMA compares most closely to auto-sklearn and TPOT as they also optimize scikit-learn pipelines. Auto-sklearn and GAMA both implement the same ensembling technique [2]. GAMA and TPOT both feature evolutionary search with NSGA2 selection [3], although GAMA’s implementation uses asynchronous evolution, which is often faster. While TPOT and auto-sklearn have a fixed AutoML pipeline, they do allow modifications to their search space. To the best of our knowledge, GAMA is the only AutoML framework that offers a modular and extensible composition of AutoML systems, and extensive support for AutoML research.

4 Conclusion

In this proposal we presented GAMA, an open-source AutoML tool which facilitates AutoML research and skillful use through its modular design and built-in logging and visualization. Novice users can make use of the graphical interface to start GAMA, or simply use the default configuration which is shown to generate models of similar performance to other AutoML frameworks. Researchers can leverage GAMA’s modularity to integrate and test new AutoML search procedures in combination with other readily available building blocks, and then log, visualize, and analyze their behavior, or run extensive benchmarks. In future work, we aim to integrate additional search techniques as well as extend the AutoML pipeline with additional steps, such as warm-starting the pipeline search with meta-data.


This software was developed with support from the Data Driven Discovery of Models (D3M) program run by DARPA and the Air Force Research Laboratory.