Inducing Interpretable Voting Classifiers without Trading Accuracy for Simplicity: Theoretical Results, Approximation Algorithms

06/09/2011
by   R. Nock, et al.
0

Recent advances in the study of voting classification algorithms have brought empirical and theoretical results clearly showing the discrimination power of ensemble classifiers. It has been previously argued that the search of this classification power in the design of the algorithms has marginalized the need to obtain interpretable classifiers. Therefore, the question of whether one might have to dispense with interpretability in order to keep classification strength is being raised in a growing number of machine learning or data mining papers. The purpose of this paper is to study both theoretically and empirically the problem. First, we provide numerous results giving insight into the hardness of the simplicity-accuracy tradeoff for voting classifiers. Then we provide an efficient "top-down and prune" induction heuristic, WIDC, mainly derived from recent results on the weak learning and boosting frameworks. It is to our knowledge the first attempt to build a voting classifier as a base formula using the weak learning framework (the one which was previously highly successful for decision tree induction), and not the strong learning framework (as usual for such classifiers with boosting-like approaches). While it uses a well-known induction scheme previously successful in other classes of concept representations, thus making it easy to implement and compare, WIDC also relies on recent or new results we give about particular cases of boosting known as partition boosting and ranking loss boosting. Experimental results on thirty-one domains, most of which readily available, tend to display the ability of WIDC to produce small, accurate, and interpretable decision committees.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/15/2011

A theory of multiclass boosting

Boosting combines weak classifiers to form highly accurate predictors. A...
research
06/16/2021

A Dataset-Level Geometric Framework for Ensemble Classifiers

Ensemble classifiers have been investigated by many in the artificial in...
research
09/10/2019

Boosting Classifiers with Noisy Inference

We present a principled framework to address resource allocation for rea...
research
02/27/2013

Induction of Selective Bayesian Classifiers

In this paper, we examine previous work on the naive Bayesian classifier...
research
08/02/2023

When Analytic Calculus Cracks AdaBoost Code

The principle of boosting in supervised learning involves combining mult...
research
05/27/2014

Layered Logic Classifiers: Exploring the `And' and `Or' Relations

Designing effective and efficient classifier for pattern analysis is a k...
research
08/31/2020

Classifier Combination Approach for Question Classification for Bengali Question Answering System

Question classification (QC) is a prime constituent of automated questio...

Please sign up or login with your details

Forgot password? Click here to reset