Model-Agnostic Approaches to Multi-Objective Simultaneous Hyperparameter Tuning and Feature Selection

12/30/2019
by   Martin Binder, et al.
0

Highly non-linear machine learning algorithms have the capacity to handle large, complex datasets. However, the predictive performance of a model usually critically depends on the choice of multiple hyperparameters. Optimizing these (often) constitutes an expensive black-box problem. Model-based optimization is one state-of-the-art method to address this problem. Furthermore, resulting models often lack interpretability, as models usually contain many active features with non-linear effects and higher-order interactions. One model-agnostic way to enhance interpretability is to enforce sparse solutions through feature selection. It is in many applications desirable to forego a small drop in performance for a substantial gain in sparseness, leading to a natural treatment of feature selection as a multi-objective optimization task. Despite the practical relevance of both hyperparameter optimization and feature selection, they are often carried out separately from each other, which is neither efficient, nor does it take possible interactions between hyperparameters and selected features into account. We present, discuss and compare two algorithmically different approaches for joint and multi-objective hyperparameter optimization and feature selection: The first uses multi-objective model-based optimization to tune a feature filter ensemble. The second is an evolutionary NSGA-II-based wrapper-approach to feature selection which incorporates specialized sampling, mutation and recombination operators for the joint decision space of included features and hyperparameter settings. We compare and discuss the approaches on a variety of benchmark tasks. While model-based optimization needs fewer objective evaluations to achieve good performance, it incurs significant overhead compared to the NSGA-II-based approach. The preferred choice depends on the cost of training the ML model on the given data.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/09/2022

Multi-objective hyperparameter optimization with performance uncertainty

The performance of any Machine Learning (ML) algorithm is impacted by th...
research
07/17/2023

Multi-Objective Optimization of Performance and Interpretability of Tabular Supervised Machine Learning Models

We present a model-agnostic framework for jointly optimizing the predict...
research
06/15/2022

Multi-Objective Hyperparameter Optimization – An Overview

Hyperparameter optimization constitutes a large part of typical modern m...
research
03/02/2023

Evolutionary Computation in Action: Feature Selection for Deep Embedding Spaces of Gigapixel Pathology Images

One of the main obstacles of adopting digital pathology is the challenge...
research
06/20/2023

GIO: Gradient Information Optimization for Training Dataset Selection

It is often advantageous to train models on a subset of the available tr...
research
08/12/2020

Implications on Feature Detection when using the Benefit-Cost Ratio

In many practical machine learning applications, there are two objective...
research
03/20/2023

A fuzzy adaptive evolutionary-based feature selection and machine learning framework for single and multi-objective body fat prediction

Predicting body fat can provide medical practitioners and users with ess...

Please sign up or login with your details

Forgot password? Click here to reset