Simplification of Forest Classifiers and Regressors

12/14/2022
by   Atsuyoshi Nakamura, et al.
0

We study the problem of sharing as many branching conditions of a given forest classifier or regressor as possible while keeping classification performance. As a constraint for preventing from accuracy degradation, we first consider the one that the decision paths of all the given feature vectors must not change. For a branching condition that a value of a certain feature is at most a given threshold, the set of values satisfying such constraint can be represented as an interval. Thus, the problem is reduced to the problem of finding the minimum set intersecting all the constraint-satisfying intervals for each set of branching conditions on the same feature. We propose an algorithm for the original problem using an algorithm solving this problem efficiently. The constraint is relaxed later to promote further sharing of branching conditions by allowing decision path change of a certain ratio of the given feature vectors or allowing a certain number of non-intersected constraint-satisfying intervals. We also extended our algorithm for both the relaxations. The effectiveness of our method is demonstrated through comprehensive experiments using 21 datasets (13 classification and 8 regression datasets in UCI machine learning repository) and 4 classifiers/regressors (random forest, extremely randomized trees, AdaBoost and gradient boosting).

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/27/2019

Best-scored Random Forest Classification

We propose an algorithm named best-scored random forest for binary class...
research
03/21/2018

Boosting Random Forests to Reduce Bias; One-Step Boosted Forest and its Variance Estimate

In this paper we propose using the principle of boosting to reduce the b...
research
02/12/2022

The Impact of Using Regression Models to Build Defect Classifiers

It is common practice to discretize continuous defect counts into defect...
research
03/06/2023

Very fast, approximate counterfactual explanations for decision forests

We consider finding a counterfactual explanation for a classification or...
research
05/24/2019

HDI-Forest: Highest Density Interval Regression Forest

By seeking the narrowest prediction intervals (PIs) that satisfy the spe...
research
04/12/2016

An incremental linear-time learning algorithm for the Optimum-Path Forest classifier

We present a classification method with incremental capabilities based o...
research
07/20/2016

Dappled tiling

We consider a certain tiling problem of a planar region in which there a...

Please sign up or login with your details

Forgot password? Click here to reset