Permutation Decision Trees

06/05/2023
by   Harikrishnan N B, et al.
0

Decision Tree is a well understood Machine Learning model that is based on minimizing impurities in the internal nodes. The most common impurity measures are Shannon entropy and Gini impurity. These impurity measures are insensitive to the order of training data and hence the final tree obtained is invariant to any permutation of the data. This leads to a serious limitation in modeling data instances that have order dependencies. In this work, we propose the use of Effort-To-Compress (ETC) - a complexity measure, for the first time, as an impurity measure. Unlike Shannon entropy and Gini impurity, structural impurity based on ETC is able to capture order dependencies in the data, thus obtaining potentially different decision trees for different permutations of the same data instances (Permutation Decision Trees). We then introduce the notion of Permutation Bagging achieved using permutation decision trees without the need for random feature selection and sub-sampling. We compare the performance of the proposed permutation bagged decision trees with Random Forests. Our model does not assume that the data instances are independent and identically distributed. Potential applications include scenarios where a temporal order present in the data instances is to be respected.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/25/2015

Unifying Decision Trees Split Criteria Using Tsallis Entropy

The construction of efficient and effective decision trees remains a key...
research
06/26/2011

Temporal Decision Trees: Model-based Diagnosis of Dynamic Systems On-Board

The automatic generation of decision trees based on off-line reasoning o...
research
11/26/2022

Mixture of Decision Trees for Interpretable Machine Learning

This work introduces a novel interpretable machine learning method calle...
research
07/24/2020

A Nonparametric Test of Dependence Based on Ensemble of Decision Trees

In this paper, a robust non-parametric measure of statistical dependence...
research
10/18/2017

On Using Linear Diophantine Equations to Tune the extent of Look Ahead while Hiding Decision Tree Rules

This paper focuses on preserving the privacy of sensitive pat-terns when...
research
12/13/2021

Tree-Based Dynamic Classifier Chains

Classifier chains are an effective technique for modeling label dependen...
research
06/20/2020

Model family selection for classification using Neural Decision Trees

Model selection consists in comparing several candidate models according...

Please sign up or login with your details

Forgot password? Click here to reset