ORSA: Outlier Robust Stacked Aggregation for Best- and Worst-Case Approximations of Ensemble Systems

11/17/2021
by   Peter Domanski, et al.
0

In recent years, the usage of ensemble learning in applications has grown significantly due to increasing computational power allowing the training of large ensembles in reasonable time frames. Many applications, e.g., malware detection, face recognition, or financial decision-making, use a finite set of learning algorithms and do aggregate them in a way that a better predictive performance is obtained than any other of the individual learning algorithms. In the field of Post-Silicon Validation for semiconductor devices (PSV), data sets are typically provided that consist of various devices like, e.g., chips of different manufacturing lines. In PSV, the task is to approximate the underlying function of the data with multiple learning algorithms, each trained on a device-specific subset, instead of improving the performance of arbitrary classifiers on the entire data set. Furthermore, the expectation is that an unknown number of subsets describe functions showing very different characteristics. Corresponding ensemble members, which are called outliers, can heavily influence the approximation. Our method aims to find a suitable approximation that is robust to outliers and represents the best or worst case in a way that will apply to as many types as possible. A 'soft-max' or 'soft-min' function is used in place of a maximum or minimum operator. A Neural Network (NN) is trained to learn this 'soft-function' in a two-stage process. First, we select a subset of ensemble members that is representative of the best or worst case. Second, we combine these members and define a weighting that uses the properties of the Local Outlier Factor (LOF) to increase the influence of non-outliers and to decrease outliers. The weighting ensures robustness to outliers and makes sure that approximations are suitable for most types.

READ FULL TEXT

page 1

page 7

research
12/13/2013

An Extensive Evaluation of Filtering Misclassified Instances in Supervised Classification Tasks

Removing or filtering outliers and mislabeled instances prior to trainin...
research
02/01/2005

Neural network ensembles: Evaluation of aggregation algorithms

Ensembles of artificial neural networks show improved generalization cap...
research
03/23/2023

Enriching Neural Network Training Dataset to Improve Worst-Case Performance Guarantees

Machine learning algorithms, especially Neural Networks (NNs), are a val...
research
07/29/2020

Outlier-Robust Estimation: Hardness, Minimally-Tuned Algorithms, and Applications

Nonlinear estimation in robotics and vision is typically plagued with ou...
research
10/22/2020

Optimal Approximation – Smoothness Tradeoffs for Soft-Max Functions

A soft-max function has two main efficiency measures: (1) approximation ...
research
06/21/2020

On Aggregation in Ensembles of Multilabel Classifiers

While a variety of ensemble methods for multilabel classification have b...
research
07/03/2018

Coopetitive Soft Gating Ensemble

In this article, we proposed the Coopetititve Soft Gating Ensemble or CS...

Please sign up or login with your details

Forgot password? Click here to reset