1 Introduction
Decision trees are traditionally built using heuristic methods, such as CART (Breiman et al. (1984)), and can produce highquality trees in low computational time. A commonly criticised point, however, is that heuristically constructed decision trees may not necessarily be the best representation of the data in terms of accuracy, size, or other considerations such as fairness.
An alternative is to construct optimal decision trees, i.e., the best possible decision tree according to a given metric. The idea of computing optimal decision trees dates back to approximately the 1970s when constructing optimal decision trees was proven to be hard by Laurent and Rivest (1976). As emphasised by Bertsimas and Dunn (2017), while optimal decision trees have always been desirable, the authors of the CART algorithm (Breiman et al. (1984)) found that such trees were computationally infeasible given the resources of the time, and hence heuristic algorithms were the only option.
Optimal decision trees are enticing for several reasons. It has been observed that a more accurate representation of the data offers better generalisation on unseen data (Bertsimas and Dunn (2017); Verwer and Zhang (2017, 2019)). This has been reiterated in our experiments as well. Optimal decision trees allow incorporating additional constraints that may be difficult to support in a heuristic algorithm. This is particularly important in sociallysensitive contexts, where special measures need to be taken to ensure fairness in machine learning. Otherwise, obtained models may implicitly or explicitly perpetuate discrimination and biases, reducing social welfare (Aghaei et al. (2019)). In some applications, the goal is to optimise the size of the decision tree representing a given controller to save memory for embedded devices (Ashok et al. (2020)). Decision trees, in particular those of small size, are desirable for formal methods when verifying properties of trained controllers, as opposed to more complex machine learning models (Bastani et al. (2018)). In recent years, there has been growing interest in
explainable artificial intelligence
. The basic premise is that machine learning models, apart from high accuracy, must also be able to explain their decisions to a (nonexpert) human. This is necessary to increase human trust and reliability of machine learning in complex scenarios that are conventionally handled by humans. Optimal decision trees of small size naturally fit within the scope of explainable AI, as their reduced size is more convenient for human interpretation.Learning problems are defined as mathematical programs: an objective function is posed possibly together with a set of constraints. An advantage of optimal decision tree algorithms over heuristic approaches is that they adhere precisely to the given specification. This allows a clear analysis and assessment of the suitability of the particular mathematical formulation for a given application. In contrast, in heuristic methods there is a discrepancy between the target learning problem and the goals of the heuristic algorithm. In more detail, heuristic methods for decision trees do not necessarily directly optimise according to the learning problem, but rather locally optimise a sequence of subproblems with respect to a surrogate metric. While this has shown to produce reasonably accurate models quickly, it may be difficult to make conclusive statements on the learning problem definition, as the heuristic approach may not faithfully follow the desired metrics. For example, a specification might be deemed suboptimal not due to a flaw in the definition, but rather because of the inability of the heuristic algorithm to optimise according to the specification.
Despite the appeal of optimal algorithms for decision trees, heuristic methods are historically the dominant approach due to computational reasons. As both algorithmic techniques and hardware advanced, optimal decision trees have become within practical reach and attracted growing interest from the research community. In particular, there has been a surge of successful methods in the past few years. These approaches use generic optimisation methods, namely integer programming (Bertsimas and Dunn (2017); Verwer and Zhang (2017, 2019); Aghaei et al. (2019)), constraint programming (Verhaeghe et al. (2019)), and SAT (Narodytska et al. (2018)), and algorithms tailored to the decision tree problem (Nijssen and Fromont (2007); Hu et al. (2019); Aglin et al. (2020)). The methods DL8 (Nijssen and Fromont (2007)) and DL8.5 (Aglin et al. (2020)) are of particular interest as they can be seen as a starting point for our work. The DL8.5 approach has been shown to be highly effective, outperforming the other approaches, and is a demonstration that specialised methods may have an advantage over generic optimisation.
Our Contribution. While previous works use highly related ideas, the presentation and terminology may differ substantially. In this work, we unify and generalise successful concepts from the literature by viewing the problem through the lens of a conventional algorithmic framework, namely dynamic programming and search. We introduce novel algorithmic techniques that reduce computation time by orders of magnitude when compared to the stateoftheart. This notably contributes towards the practical application of optimal classification trees, which was traditionally plagued by high runtimes. We conduct an experimental study on a wide range of benchmarks from the literature to show the effectiveness of our approach and its components, and reiterate that optimal decision trees lead to better generalisation in terms of outofsample accuracy. Our framework supports constraints on the depth of the tree and the number of nodes, and we argue it is flexible and may be extended with other requirements. In more detail, the contributions are as follows:

MurTree (Section 4), a framework for computing optimal classification trees, i.e., decision trees that minimise the number of misclassifications. The framework allows constraints on the depth and the number of nodes of the decision tree. The node constraint is not considered in all works on optimal decision trees, notably it is not supported by the previously fastest algorithm, DL8.5. Additional objective functions and constraints may be added that admit a dynamic programming formulation (Section 4.9).

A clear highlevel view of the framework using conventional algorithmic principles, namely dynamic programming and search, that unifies and generalises ideas from the literature (Section 4.1).

A specialised algorithm for computing the optimal classification tree of depth two, which serves as the backbone of our framework (Section 4.3). It uses a frequency counting method to avoid explicitly referring to the dataset. This substantially reduces the runtime of computing optimal trees which, when combined with an incremental technique that takes into account previous computations, provides orders of magnitude speedups.

A novel similaritybased lower bound on the number of misclassifications for an optimal decision tree. The bound is effective in determining that portions of the search space cannot contain better decision trees than currently found during the search, which allows the algorithm to prune parts of the search space without needing further inspection, providing additional speedups. The bound is derived by examining previously computed subtrees and computing the number of misclassifications that must hold in the new search space (Section 4.5).

We incorporate the constraint on the number of nodes in the tree, extend the caching technique to take into account both the depth and number of nodes constraint (Section 4.6), refine the lower bounding technique on the number of misclassifications from DL8.5 (Aglin et al. (2020)) to produce stronger bounds (4.6.1), and provide an incremental solving option to allow reusing computations when solving a series of increasingly large decision trees (Section 4.6.3), e.g., as encountered in hyperparameter tuning. Further improvements include a dynamic postorder node exploration strategy (Section 4.7) that leads to consistent improvements over a conventional postorder search.

We provide a detailed experimental study to analyse the effectiveness of our individual techniques and scalability of our approach, evaluate our approach with respect to the stateoftheart optimal classification tree algorithms, and compare against heuristic decision tree and random forest algorithms on outofsample accuracy (Section
5). The experimental results show that our approach provides highly accurate trees and exhibits speedups of (several) orders of magnitude when compared to the stateoftheart.
The rest of the paper is organised as follows. In the next section, we introduce the notions and definitions used throughout the paper. In Section 3, we review the stateoftheart for optimal decision trees. Our main contribution is given in Section 4, where we describe our MurTree framework. In Section 5, we conduct a series of empirical evaluations of our approach and conclude in Section 6.
2 Preliminaries
A feature is a variable that encodes information about an object. We speak of binary , categorical , and continuous features depending on their domain, i.e., , , and . A feature vector
is a vector of features. An
instance is a pair that consists of a feature vector and a value representing the class. A class can take continuous or discrete values. A dataset, or simply data, is a set of instances. While features within a vector may have different domains, the ith feature of each feature vector of the dataset shares the same domain. The assumption is that the features describe certain characteristics about the objects, and the ith feature of each feature vector refers to the same characteristic of interest.The process of learning seeks to compute a learning function that performs classification, i.e., maps feature vectors to classes. The target learning function is restricted to a particular form, e.g., the form of a decision tree (see further), and the goal is to compute a function that minimises or maximises the target metric for a given dataset. If the domain of the classes of the dataset is discrete, we speak of a classification problem, and otherwise of a regression problem for continuous classes.
Decision trees are binary trees from computer science. We call leaf and nonleaf nodes classification and predicate nodes, respectively. Each predicate node is given a predicate that maps feature vectors to a Boolean value, i.e., . The left and right edges of a predicate node are associated with the values zero and one, respectively. Each classification node is assigned a fixed class.
A decision tree is a learning function that performs classification according to the following recursive procedure. Given a feature vector, it starts by considering the root node. If the considered node is a classification node, its class determines the class of the feature vector and the procedure terminates. Otherwise, the node is a predicate node, and the left child node will be considered next if the predicate of the node evaluates to zero, and otherwise the right child node is selected. The process recurses until a class is determined.
The (feature) depth of a decision tree is the maximum number of feature nodes any instance may encounter during classification. The size of a decision tree is the number of feature nodes. It follows that the maximum size of a decision tree with depth is . We note that in the literature, in some cases, the size is defined as the total number of nodes in the tree. These definitions are equivalent and can be used interchangeably, as a tree with predicate nodes has classification nodes.
In practice, the predicates take a special form. For singlevariate or axisaligned decision trees, which are the focus of this work, predicates only consider a single feature and typically test whether it exceeds a threshold value. We refer to these nodes are feature nodes, as the predicate depends solely on one feature. Furthermore, the predicates are chosen based on the dataset. Generalisations of decision trees are straightforward: multivariate versions use predicates that operate on more than one feature, and predicates can be substituted by functions whose codomains are of size , in which case the decision tree is an ary tree with an analogous definition. These generalisations are mentioned for completeness and are not further discussed.
We use special notation for binary datasets, where the domain of features and classes is Boolean. Given a feature vector , we write and if the ith feature has value one and zero, respectively. The value one indicates the feature is present in the feature vector, and otherwise it is not present. Features and are referred to as positive and negative features, respectively. We limit the predicates to only output the value of a particular feature in the feature vector and simply write and for the predicates. The binary dataset is partitioned into a positive and negative class of instances based on the classes, i.e., . We consider the partitions as sets of feature vectors since their class is clear from context, and write as the set of instances from that contain feature , and analogously for multiple features, e.g., are the set of instances that contain both and . The misclassification score of a decision tree on data is the number of instances for which classification produces the incorrect class considering the data as ground truth.
3 Literature Review
Historically the most popular techniques for decision tree learning were based on heuristics due to their effectiveness and scalability. Examples of these algorithms include CART, originally proposed by Breiman et al. (1984), and C4.5 by Quinlan (1993). These algorithms start with a single node, and iteratively expand the tree based on metrics such as information gain and Gini coefficients, and possibly postprocess the obtained decision trees to prune branches in an effort to reduce overfitting. While there is a vast literature on heuristic algorithms for decision trees, in this work we are primarily concerned with optimal decision trees, and hence direct further discussion to such settings.
Bertsimas and Shioda (2007) presented a mixedinteger programming approach for optimal decisions that worked well on smaller datasets. Mixedinteger programming formulations with better performance were given by Bertsimas and Dunn (2017) and Verwer and Zhang (2017). These methods encode the optimal decision tree by fixing the tree depth in advance, creating variables to represent the predicates for each node, and adding constraints to enforce the decision tree structure. These approaches were later improved by BinOPT (Verwer and Zhang (2019)), a
binary linear programming
formulation, that took advantage of implicitly binarising data to reduce the number of variables and constraints required to encode the problem. Aghaei et al. (2019) used a mixedinteger programming formulation for optimal decision trees that supported fairness metrics. The authors argued that using machine learning in socially sensitive contexts may perpetuate discrimination if no special measures are taken into account. They propose fairness metrics and incorporate them in a mixedinteger programming formulation.An encoding of decision trees using propositional logic (SAT) has been devised by Narodytska et al. (2018). In this line of work, the aim is to construct the smallest tree in terms of the total number of nodes that perfectly describes the given dataset, i.e., leads to zero misclassifications on the training data. An initial perfect decision tree is constructed using a heuristic method, after which a series of SATsolver calls are made, each time posing the problem of computing a perfect tree with one less node. The SAT approach of Avellaneda (2020) simplifies the encoding by fixing the depth of the tree and employs an incremental approach where instances are gradually added to the formulation rather than being considered completely from the start.
Nijssen and Fromont (2007) introduced a framework named DL8 for optimal decision trees that could support a wide range of constraints. They took advantage that the left and right subtree of a given node can be optimised independently, introduced a caching technique to save subtrees computed during the algorithm in order to reuse them at a later stage, and combined these with ideas from the pattern mining literature to compute optimal decision trees. DL8 laid an important foundation for optimal decision tree algorithms that follow.
Verhaeghe et al. (2019) approached the optimal classification tree problem by minimising the misclassifications using constraint programming. The independence of the left and right subtrees from Nijssen and Fromont (2007) was captured in an ANDOR search framework. Upper bounding on the number of misclassifications was used to prune parts of the search space and their algorithm incorporated an itemset mining technique to speedup the computation of instances per node and used a caching technique similar to DL8 (Nijssen and Fromont (2007)),
Hu et al. (2019) presented an algorithm that computes the optimal decision tree by considering a balance between misclassifications and number of nodes. They apply exhaustive search, caching, and lower bounding of the misclassifications based on the cost of adding a new node to the decision tree. Compared to other recent optimal decision tree algorithms, the method relies on the number of nodes playing an important role in the metric of optimality and a limited number of binary features, e.g., the authors experimented with datasets with up to twelve binary features.
Aglin et al. (2020) developed DL8.5 by combining and refining the ideas from DL8 and the constraint programming approach. Their main addition was an upper bounding technique, which limited the upper misclassification value of a child node once the optimal subtree was computed for its sibling, and a lowering bound technique, where the algorithm stored information not only about computed optimal subtrees but also pruned subtrees to provide a lower bound on the misclassifications of a subtree. This led to an algorithm that outperformed previous approaches by a notable margin.
Exploiting properties specific to the decision tree learning problem proved to be valuable in improving algorithmic performance in previous work. In particular, search and pruning techniques, caching computation for later reuse, and the techniques that take advantage of the decision tree structure all lead to notable gains in performance. These are the main reasons for the success of specialised methods over generic frameworks, such as integer programming and SAT. As there is a significant overlap of ideas and techniques used in related work, we discuss these in more detail in Section 4.1 when presenting the highlevel view of our framework.
Lastly, we refer the readers to a curated list of decision tree papers by Benedek Rozemberczki: https://github.com/benedekrozemberczki/awesomedecisiontreepapers.
4 MurTree: Our Framework for Optimal Classification Trees
Our framework computes optimal classification trees by exhaustive search. The search space is exponentially large, but special measures are taken to efficiently iterate through solutions, exploit the overlap between solutions, and avoid computing suboptimal decision trees.
We give the main idea of the algorithm, then provide the full pseudo code, and follow up with individual subsections where we present each individual technique in greater detail.
For the sake of clarity, the remaining text focusses on optimal classification trees that minimise the number of misclassified instances for binary datasets and binary classification. Extending the framework for general settings, such as continuous and categorical data, is discussed in Section 4.9.
4.1 HighLevel Idea
We note two important properties of decision trees:
Property 1
(Independence) Given a dataset , a feature node partitions the dataset into its left and right subtree, such that and .
Property 2
(Overlap) Given a classification node, a set of features encountered on the path from the root node to the classification node, and an instance, the order in which the features are used to evaluate the instance does not change the classification result.
Both properties follow directly from the definition of decision trees and are emphasised as they play a major role in designing decision tree algorithms. Property 1 allows computing the misclassification score of the tree as the sum of the misclassification scores of its left and right subtree, and as will be discussed, once a feature node is selected, the left and right subtrees can be optimised independently of each other. Property 2 shows there is an overlap between decision trees that share the same features, which is taken advantage of by caching techniques (see Section 4.6 for more details).
The dynamic programming formulation of optimal classification trees given in Eq. 1 provides a highlevel summary of our framework. The input parameters consist of a binary dataset with features , an upper bound on depth , and an upper bound on the number of feature nodes . The output is the minimum number of misclassifications possible on the data given the input decision tree characteristics. The key observations are given by Properties 1 and 2. The two observations, independence and overlap, when combined reveal the dynamic programming structure of decision trees.
(1) 
The first case in Eq. 1 places a natural limit on the number of feature nodes given the depth. The second case defines the misclassification score for classification nodes. The general case states that computing the optimal misclassification score amounts to examining all possible feature splits and ways to distribute the feature node count to the left and right child of the root node. For each combination of a selected feature and node count distribution to its children, the optimal misclassification is computed recursively as the sum of the optimal misclassifications of its children. The formulation is exponential in the depth, feature node limit, and number of features, but with special care, as presented in the subsequent sections, it is possible to compute practically relevant optimal classification trees within a reasonable time.
Eq. 1 serves as the core foundation of our framework. In contrast to related work, we take advantage of the structure of decision trees to allow imposing a limit on the number of nodes as presented in Eq. 1. Previous approaches either place no constraint on the number of nodes apart from the depth (Nijssen and Fromont (2007); Aglin et al. (2020)), limit the number of nodes by penalising the objective function for each node in the tree (Hu et al. (2019)), or allow constraints on the number of nodes but do not make use of decision tree properties (Bertsimas and Dunn (2017); Narodytska et al. (2018); Verwer and Zhang (2017, 2019); Avellaneda (2020)). The last point is particularly important as the ability to exploit optimal decision tree properties is essential for achieving the best performance.
Simpler and/or modified forms of Eq. 1 were used in some previous work under different terminology. The ANDOR search method (Verhaeghe et al. (2019)), pattern mining approach (Nijssen and Fromont (2007); Aglin et al. (2020)), and the search by Hu et al. (2019) use the independence property of the left and right subtree (Property 1). Those approaches save computed optimal subtrees (Property 2), which corresponds to memoisation as an integral part of dynamic programming (Section 4.6). Framing the problem as a dynamic program dates from the 1970s (e.g., Garey (1972)), but the description in works afterwards deviated as new techniques were introduced. We present the problem back in its original dynamic programming format and together with our node limitation addition, unite and generalise previous approaches using conventional algorithmic notation.
A key component of our framework is a specialised algorithm for computing decision trees of depth at most two. It takes advantage of the specific decision tree structure by performing a precomputation on the data, which allows it to compute the optimal decision tree without explicitly referring to the data. This offers a significantly lower computational complexity compared to the generic case of Eq. 1, but is applicable in practice only to decision trees of depth two. Thus, rather than following Eq. 1 until the base case, we stop the recursion once a tree of depth two is required and invoke the specialised method.
A defining characteristic of search algorithms are pruning techniques, which detect areas of the search that may be discarded without losing optimality. In the case of decision trees, subtrees may be pruned based on the lower or upper bound of the number of misclassifications of the given subtrees. If the bound shows that the misclassifications of a currently considered subtree will result in a high value, the subtree can be pruned, effectively reducing the search space. The challenge when designing bounding techniques is to find the correct balance between pruning power and the computational time required by the technique.
We introduce a novel similaritybased lower bounding technique (Section 4.5) that derives a bound based on the similarity of the previously considered subtrees. We use our lower bounding method in combination with the previous lower bounding approach introduced in DL8.5 (Aglin et al. (2020)), which we describe in the following text. Given a parent node, once the optimal subtree is computed for one of the children, an upper bound can be posed on the other child subtree based on the best decision tree known for the parent node and the number of misclassifications of the optimal child subtree. If a subtree fails to produce a solution within the posed upper bound, the upper bound is effectively a lower bound that can be used once the same subtree is encountered again in the search. Our algorithms uses a refinement of the described lower bound, which additionally takes into account all lower bounds of the children of the parent node (Section 4.6.1). Hu et al. (2019) uses a bound for an objective function that balances the accuracy (misclassifications) and number of nodes in the tree. If is the penalty in terms of misclassifications for adding a node to the decision tree, then also serves as a lower bound for each subtree (otherwise it is not worth introducing a node). We do not incorporate this last bound explicitly in our framework, but instead compute trees with such objective functions by solving a series of (overlapping) trees that optimise only the misclassification score (Section 4.9.3).
The remaining part of the paper describes our techniques in more detail.
4.2 Main Loop of the Framework
Algorithm 1 summarises our framework. As discussed in the previous section, it can be seen as an instantiation of Eq. 1 with additional techniques to speedup the computation.
The algorithm takes as input a dataset consisting of positive and negative instances, the maximum depth and size (number of feature nodes) of the decision tree, and an upper bound that represents a limit on the number of misclassifications before the tree is considered infeasible. The output is an optimal classification tree respecting the input constraints on the depth, size, and upper bound, or a flag indicating that no such tree exists, i.e., the problem is infeasible. The latter occurs as a result of recursive calls (see further), which pose an upper bound that is necessary to ensure the decision tree has a lower misclassification value than the best tree found so far in the search.
The upper bound is initially set to the misclassification score of a single classification node for the data and is updated throughout the execution. Note that at the start of the algorithm a tighter upper bound could be computed by using a heuristic algorithm.
The base case of Eq. 1 is initially tested and a classification node is returned if no feature nodes are allowed. In addition, subtrees that are at their lower bound misclassification values are already optimal and are returned immediately.
After the initial tests, the algorithm attempts to prune the current tree based on two lower bounds: our novel similaritybased approach (Section 4.5), and our generalisation of the cachebased lower bounding introduced in DL8.5 (Aglin et al. (2020)) but now extended to take into account the number of nodes in the tree.
Assuming pruning did not take place, if the current subtree has already been computed as part of a previous recursive call, the solution is retrieved from the cache and the current call terminates. Caching subtrees for trees where the depth is constrained dates from DL8 (Nijssen and Fromont (2007)). In our work, the algorithm caches with respect to the depth and number of node constraints.
A key aspect of our framework is that trees of depth at most two are computed using a specialised procedure (Section 4.3). It solves the optimal decision tree problem in a sense as a unit operation and ignores the upper bound. The result is stored in the cache for future computation regardless of the feasibility of tree with respect to the upper bound, but the upper bound determines if the obtained tree is considered feasible.
If none of the above criteria is met, the algorithm reaches the general case from Eq. 1, where the search space is exhaustively explored through a series of overlapping recursions (Algorithm 2).
Recall that the size of tree, i.e., the number of feature nodes, is given as input. One node is allocated as the root, and the remaining node budget is split among its children. For each feature, the algorithm considers all possible combinations of distributing the remaining node budget to its left and right subtrees. Note that determining the maximum size of thfe left subtree immediately fixes the maximum size of the right subset, and that special care needs to be taken to not allocate a size to a subtree that is greater than it may support with respect to its depth.
For a chosen tree configuration (the feature of the subtree root and the size of its subtrees), the algorithm determines which subtree to recurse on first. Previous work in DL8 and DL8.5 fixed the order by exploring the left before the right subtree. In our framework, we introduce a dynamic strategy that prioritises the subtree with the largest gap between its lower and upper bound (Section 4.7
). The intuition is that this subtree is more probable to have a higher misclassification score, which in turn increases the likelihood of pruning the other sibling.
The algorithm then solves the subtrees in the chosen order. If the first subtree is infeasible, this implies that the lower bound of the subtree is one greater than the given upper bound. The information is stored in the cache in case the bound is needed in one of the other recursive calls. This bound was introduced in DL8.5 Aglin et al. (2020) and we provide a further refinement by into account pairwise sum of the lower bounds of the children (Section 4.6.1). Recall that the misclassification score of the root is the sum of the misclassifications of its children, and therefore the second subtree can be discarded if its sibling already led to an infeasible tree.
If both recursive calls successfully terminated, the obtained decision tree is recorded as the best tree found so far and the solution is stored in the cache. In our framework, as soon as a new globally optimal decision tree is encountered, it is identified as such. This leads to fully anytime behaviour, i.e., the execution can be stopped at any given point in time to return its current best solution. In the previous work of DL8.5, for instance, a globally improving solution was only detected at the root node of the complete decision tree.
Once all the recursive calls have been completed, the search space of the subtree has been exhaustively explored. The cache is updated with respect to the best locally found subtree: either the subtree is stored in the cache as an optimal subtree, or its lower bound is updated in case no feasible subtree was found. In this manner, all possible decision trees are explored and a tree with minimum misclassification score is returned.
The dynamic programming aspect can be seen as the method divides the main problem into smaller overlapping subproblems, owing to Properties 1 and 2. Search is used to prune the search space, saving computation time, and drives the algorithm towards the specialised algorithm, which efficiently computes optimal subtrees of depth at most two. The last point is a key component in reducing the overall runtime compared to previous approaches, as discussed in Section 4.3.
Lastly, we note two points not included in the pseudocode for simplicity. Note the following definition and proposition:
(Degenerate Decision Trees) A decision tree is degenerate if it contains at least one classification node that does not classify any training instance.
(Pruning Degenerate Trees) Given a degenerate decision tree with feature nodes and misclassification score on the training data, there exists at least one other decision tree with feature nodes and misclassification score .
Degenerate trees may occur during the algorithm when splitting on a nondiscriminative feature, such that one subtree contains no training instances, i.e., . Due to Proposition 4.2, we deem these trees infeasible and prune them as soon as they are detected.
The second point is that the initial best subtree is set to a classification node if allowed by the upper bound, rather than an infeasible tree as given in Algorithm 1, which may trigger a global update of the best solution.
This concludes the description of the main loop of our framework. Before proceeding with detailing each component of our algorithm, we reiterate the differences between our approach and DL8.5 (Aglin et al. (2020)) in light of the technical description given above.
Comparison with DL8.5 (Aglin et al. (2020)). By virtue of taking into account the structure of decision trees, Algorithm 1 shares a similar layout as in DL8.5, but there are notable differences that result in orders of magnitude speedups. The differences can be summarised as follows: 1) we allow constraining the size of tree in addition to the depth, which is important in obtaining the smallest optimal decision, e.g., to improve interpretability or learn trees that generalise on unseen instances (Section 5.4), 2) our specialised algorithm (Section 4.3) is substantially more efficient at computing trees with depth two when compared to the general algorithm in Algorithm 1 or DL8.5, 3) we propose a new lower bound based on the similiarity with previously computed subtrees to further prune the search space (Section 4.5) and refine the previous lower bound (4.6.1), 4) our cache policy (Section 4.6) is extended to support the size of the tree constraint and allows for incremental solving, allowing reusing computation when solving trees with increasing depth and size, e.g., during hyperparameter tuning, 5) we dynamically determine which subtree to explore first based on pruning potential (Section 4.7), rather than use a static stategy, and 6) our framework immediately updates the best global solution as soon as it is computed rather than only at the root node.
4.3 Specialised Algorithm for Trees of Depth Two
An essential part of our framework is a specialised method for computing optimal decision trees of depth two. The procedure is repeatedly called in our framework, i.e., each time a tree of at most depth two needs to be optimally solved. In the following, we present an algorithm that achieves lower complexity than the general algorithm (Eq. 1 and Prop. 4.3) when considering trees with depth two.
Prior to presenting our specialised algorithm, we discuss the complexity of computing decision trees of depth two using Eq. 1 as the baseline.
Computing the optimal classification tree of depth two using Eq. 1 can be done in time.
Assume that splitting the data based on a feature node is done in time. Eq. 1 considers splits for root and for each feature performs splits for its children. This results in splits and an overall runtime of , proving Proposition 4.3. In practice, partitioning the dataset based on a feature can be spedup using bitvector operations and caching subproblems (Aglin et al. (2020); Verhaeghe et al. (2019); Hu et al. (2019)), but the complexity remains as this only impacts the hidden constant in the bigO.
In the following, we present an algorithm with lower complexity and additional practical improvements which, when combined, reduce the runtime of computing the optimal classification tree of depth two by orders of magnitudes.
4.4 Algorithm Description
Algorithm 3 provides a summary. The input is a dataset and the output is the optimal classification tree of depth two with three feature nodes that minimises the number of misclassified instances.
The specialised procedure computes the optimal decision tree in two phases. In the first step, it computes frequency counts for each pair of features, i.e., the number of instances in which both features are present. In the second step, it exploits the frequency counts to efficiently enumerate decision trees without needing to explicitly refer to the data. This provides a substantial speedup compared to iterating through features and splitting data as given in the dynamic programming formulation (Eq. 1) for decision trees of depth two. We now discuss each phase in more detail and present a technique to incrementally compute the frequency counts.
4.4.1 Phase One: Frequency counting (Algorithm 3, Lines 33)
Let and denote the frequency counts in the positive instances for a single feature and a pair of features, respectively. The functions and are defined analogously for the negative instances.
A key observation is that based on and , we may compute , , , and . This is done as follows:
(2) 
(3) 
(4) 
(5) 
The equations make use of the fact that the features are binary. For example, Eq. 2 states that if the total number of positive instances is and we computed the frequency count , then the frequency count is the number of instances in which does not appear, i.e., the difference between and . Similar reasoning is applied to the other equations and computing the frequency count is analogous.
The following proposition summarises the runtime of computing .
(Computational Complexity of Phase One) Let denote the maximum number of features in any single positive instance. Frequency counts can be computed in time with memory.
An efficient way of computing the frequency counts is to represent the feature vector as a sparse vector, and iterate through each instance in the dataset and increase a counter for each individual feature and each pair of features. This leads to the proposed complexity result. The additional memory is required to store the frequency counters, allowing to query a frequency count as a constant time operation. Note that the pairwise frequency count is symmetric, i.e., , which requires only to consider and in the frequency count for . This results in a smaller hidden constant in the bigO notation.
4.4.2 Phase Two: Optimal tree computation (Algorithm 3, Lines 33)
Recall that a classification node is assigned the positive class if the number of positive instances exceeds the number of negative instances, otherwise the node class is negative. Let be the classification score for a classification node with all instances of containing both features and . The classification score is then computed as follows.
(6) 
Given a decision tree with depth two, a root node with feature , a left and right child node with features and , we may compute the misclassification score in constant time assuming the frequency counts are available. Let and denote the misclassification scores of the left or right subtree. The computations are as follows.
(7) 
(8) 
The total misclassification score of the tree is the sum of misclassifications of its children. As the number of misclassification can be computed solely based on the frequency counts, we may conclude the computational complexity.
(Computational Complexity of Phase Two) Given the frequency counts and , the optimal subtree tree can be computed in time with memory.
It follows from Property 1 that given a root node with feature , the left and right subtrees can be optimised independently. Therefore, it is sufficient to compute for each feature its best left and right subtrees, and take the feature with the minimum sum of its child misclassifications. To compute the best left and right feature for each feature, the algorithm maintains information about the best left and right child for each feature found so far, leading to the memory requirement from Proposition 4.4.2. The best features are initially arbitrarily chosen. Recall that from Property 1 it follows that the left and right subtree can be optimised independently:
Therefore, rather than considering triplets of features , it iterates through each pair of features , computes the misclassification values of the left subtree using Eq. 7, updates the best left child for feature , and performs the same procedure for the right child. After iterating through all pairs of features, the best left and right subtree is known for each feature, leading to the proposed complexity. The optimal decision tree can then be computed by finding the feature with minimum misclassification cost of its combined left and right misclassification.
After discussing each individual phase, we may conclude the overall complexity:
(Computational Complexity of Depth2 Decision Trees) Let be the upper limit on the number of features in any single positive and negative instance. The number of operations required to computing an optimal decision tree is using auxiliary memory.
The result follows by combining Propositions 4.4.1 and 4.4.2. The obtained runtime is substantially lower at the expense of using additional memory compared to the dynamic programming formulation (Eq. 1) outlined in Proposition 4.3. Note that instances with binary features are naturally sparse. If the majority of instances contain more than half of the features, then as a preprocessing step all feature values may be inverted to achieve sparsity without loss of generality. The advantage of our approach is exemplified with lower sparsity ratios, i.e., small values.
There are several additional points to note, which are not shown in Algorithm 3 to keep the pseudocode succinct.
The above discussion assumed the feature node limit was set to three. The algorithm can be modified for the case of two feature nodes, keeping the same complexity, while in the case with only one feature node the pairwise computations are no longer necessary leading to complexity. Similarly, the algorithm is implemented to lexicographically minimise the misclassification score and then the size of the tree.
To improve the performance in practice, the algorithm iterates through pairs of features such that . After updating the current best left and right subtree feature using as the root and as the child, the same computation is done using as the root and as the child. Compared to the pseudocode in Algorithm 3, this cuts the number of iterations by half, but each iteration does twice as much work, which results in a speedup in practice. Moreover, rather than computing the best tree in a separate loop after computing the best left and right subtrees for each feature, this is done on the fly by keeping track of the best subtree encountered so far during the algorithm.
Specialised algorithm for decision trees of depth three. We considered computing decision trees with depth three using a similar idea. Even though this results in a better bigO complexity for trees of depth three, albeit requiring memory, our preliminary results did not indicate practical benefits. Including additional lowlevel optimisation might improve the results, but for the time being we leave this as an open question.
4.4.3 Incremental Computation
The specialised method for computing decision trees of depth two is repeatedly called in the framework. For each call, the algorithm is given a different dataset that is a result of applying a split in one of the nodes in the tree. The key observation is that datasets which differ only in a small number of instances result in similar frequency counts. The idea is to exploit this by only updating the necessary difference rather than recomputing the frequency counts from scratch.
The key point is to view the previous dataset and the new dataset in terms of their intersection and differences.
Observation 1
Given two datasets and , let their difference be denoted as and and their intersection as . We may express the datasets as and
We first note that set operations can be done efficiently for datasets.
(Computational Complexity of Set Operations on Datasets) Given a dataset and two of its subsets and , the sets and can be computed in time using memory.
The above can be realised by associating each instance of the original dataset with a unique ID and afterwards using direct hashing to query in constant time the presence of an instance in a dataset. Once the differences have been computed, the frequency counts may be updated incrementally.
(Computational Complexity of Incremental Frequency Computation) Let denote the maximum number of features in any considered instance. Given the frequency counts of a previous dataset , a new dataset , and their differences and , the frequency counts of the new dataset can be computed in time.
To show the complexity, note the difference between and .
Observation 2
Let denote the set of instances used to compute the frequency counts . It follows that and .
Consider taking and applying a series of operations to reach the new frequency counts . The complexity result of Proposition 1 follows from the previous observations and the following:
Observation 3
The frequency counts already capture the counts for instances
Observation 4
The frequency counts need to be incremented using instances
Observation 5
The frequency counts need to be decremented using instances
Using the incremental update procedure is sensible only if the number of updates required is small compared to recomputing from scratch. Therefore, in our framework, in each call to compute a decision tree of depth two, the algorithm incurs an overhead (Proposition 1) to compute the differences between the old and new dataset. It proceeds with the incremental computation if , and otherwise computes from scratch.
Note that the overhead is negligible compared to the overall complexity of computing the optimal tree of depth two (Proposition 4.4.2), but the benefits can be significant if the difference is small. As shown in the experimental section, this is frequently the case in practice, as two successive features considered for splitting are unlikely to lead to vastly different splits.
4.5 SimiliarityBased Lower Bounding
We present a novel lower bounding technique that does not rely on the algorithm having previously searched a specific path, as opposed to the cachebased lower bound introduced in the later sections. Given a dataset for a node, the method aims to derive a lower bound by taking into account a previously computed optimal decision tree using the dataset . It infers the bound by considering the difference in the number of instances between the previous dataset and the current dataset . The bound is used to prune portions of the search space that are guaranteed to not contain a better solution than the best decision tree encountered so far in the search.
Assume that for both datasets, the depth and the number of allowed feature nodes requirements are identical. As in the previous section, we define the sets , , and .
Given the limits on the depth and number of features nodes , a dataset , and a dataset with as the misclassification score of the optimal decision tree of (recall Eq. 1), we define the similaritybased lower bound,
(9) 
which is a lower bound for the number of misclassifications of the optimal decision tree for the dataset of a tree of depth with feature nodes, i.e.,
.
As a result, subtrees with a lower bound greater than its upper bound are immediately pruned, effectively speeding up the search. To show that Proposition 4.5 is indeed a lower bound, let , note that removing from may reduce the misclassification cost by at most :
(10) 
(11) 
Adding instances to cannot decrease the misclassification score :
(12) 
(13) 
(14) 
which shows the derivation of Proposition 4.5.
As shown in the experimental results (Section 5.2.1), the use of the similaritybased lower bound reduces the runtime for all datasets, with only a few exceptions, and in some cases the obtained reduction is an order of magnitude.
4.6 Caching of Optimal Subtrees (Memoisation)
As is common in dynamic programming algorithms, a caching or memoisation table is maintained to avoid recomputing subproblems. In our case, information about computed optimal subtrees is stored. This is used to retrieve a subtree that has already been computed when needed, provide lower bounds, and reconstruct the optimal decision tree at the end of the algorithm. Caching has been used in previous works (Nijssen and Fromont (2007); Aglin et al. (2020); Verhaeghe et al. (2019); Hu et al. (2019)), and here we extend it to support constraints on the number of nodes and incremental solving.
The key observation is that given a path from the root to any given node, each permutation of the feature nodes on the path results in the same dataset for the node furthest from the root, e.g., . This allows representing a path as a set of features, e.g., . Each time an optimal subtree is computed during search, its path and root node are stored in the cache. If no subtree could be computed within the specified upper bound, a lower bound is derived based on the information collected during the search and the given upper bound. The bound is stored in the cache for reuse later on. Caching and lower bounds derived in this manner have been used in DL8.5 Aglin et al. (2020) to compute optimal decision trees with a given depth.
We generalise caching to support being used when the number of nodes is also a constraint, in addition to the depth, and allow incremental solving, i.e., the scenario when progressively large trees in terms of depth and size are computed in succession, e.g., during hyperparameter tuning. Furthermore, we strengthen the bound introduced in DL8.5 (Aglin et al. (2020)) by using information obtained during the search.
Recall that only the root node is stored for a given path. When necessary, the complete subtree may be reconstructed as a series of queries to the cache, where each time a single node is retrieved, as introduced in DL8 (Nijssen and Fromont (2007)). In our framework, there is an exception to the mentioned tree reconstruction procedure. After solving a tree of depth two, none of its children are stored in the cache. During the algorithm these are not necessary, but the children are needed when reconstructing the best decision tree found at the end. In this case, the required child nodes are recomputed using Algorithm 3. The computational overhead is negligible compared to the overall execution time, but this avoids storing an exponential number of paths (recall that the number of paths increases exponentially with the depth) which do not serve a purpose other than the final reconstruction.
4.6.1 Storing Subtrees and Lower Bounds in the Cache
Each node is associated with a path, represented as a set of features. Our cache maps a path to a list of cache entries, where each entry is composed of an optimal assignment, a lower bound, and the size limit. The depth is not explicitly stored as it can be derived based on the maximum depth and the number of features in the path. It is possible to store a lower bound without the optimal assignment, but note that the optimal assignment is the tightest lower bound. Initially, the cache is empty, and the lists and their entries are created dynamically during search as needed. There are two types of scenarios that prompt a cache storage.
Scenario one: node has been exhaustively searched and a solution has been found within the upper bound. In this situation, the computed subtree is optimal and the corresponding entry is stored using its root node assignment as the optimal assignment, the lower bound is set to the misclassification score, and the feature node limit is the limit that was assigned to the node.
In case the algorithm determines that the lowest classification score may be achieved using fewer nodes than imposed by the node limit, we may use the following proposition to create additional cache entries:
Let be the misclassification score of the optimal decision tree for the dataset with depth limit and node limit . If there exists an such that , then for .
During the algorithm, a given path can only be exhaustively explored once, as the next time the path is encountered its corresponding solution is retrieved from the cache (Section 4.6.2).
Scenario two: a node has been exhaustively explored, but no solution has been found within the upper bound. It follows that the optimal subtree corresponding to the path has at least as many misclassifications as the upper bound incremented by one. This is the lower bounding reasoning introduced in DL8.5 (Aglin et al. (2020)).
In this work, we propose a stronger lower bound. Let be the lower bound for the number of misclassifications of an optimal decision tree for the dataset with n nodes and depth d, i.e., . We introduce the following refined lower bound :
(15) 
The bound considers all possible assignments of features and sizes to the root and its children, and selects the minimum sum of the lower bounds of its children. It follows that no decision tree may have a misclassification score lower than . We combine with the upper bound to obtain a lower bound for the case where no decision tree with less than the specified upper bound could be found:
(16) 
The proposed bound generalises the bound from DL8.5 (Aglin et al. (2020)), which only considers the second expression on the righthand side of Eq. 16 to derive a lower bound when no tree could be found within the given upper bound. In our experiments, we observed the strengthened bound provides a speedup by a factor of at most two on several datasets.
Once the bound has been computed, it is recorded in a cache entry for the path, along with the size limit of the node, and the optimal assignment is set to null.
4.6.2 Retrieving Subtrees and Lower Bounds from the Cache
When a new child node is created, the algorithm searches through the cache to detect if the optimal solution is already present. Ideally, the cache entry of the path matches the size limit imposed on the node, but a lower bound for the current tree may be inferred from the bounds of the larger tree, formally summarised in the following proposition.
Given the dataset and depth bound and the maximum number of feature nodes , a bound for a larger tree is a bound for the current tree, i.e., .
When retrieving a lower bound and no lower bound has been stored for the currently queried decision tree, Proposition 4.6.2 allows inferring a lower bound from larger trees that may be stored in the cache. Note that the lower bounds are nonincreasing with size, i.e.,
(17) 
The tightest bound is returned when retrieving the lower bound. Assuming no bound with the prescribed number of nodes is stored, the bound of the smallest tree that is greater than the target size is returned should such an entry exist. If there are no applicable entries in the cache, the trivial lower bound of zero is returned. For example, given a dataset with the node limit set to five, if there is no subtree for the given size in the cache but there is an entry when the node limit was set to six and seven, then the lower bound using six nodes is the tightest valid lower bound available for the tree with five nodes.
4.6.3 Incremental Solving
We label incremental solving as the process of querying the algorithm to compute progressively larger decision trees. For example, once the algorithm has computed an optimal decision tree for a given depth and size, a user may be interested in a tree with more nodes to understand if the additional nodes would lead to a meaningful decrease in the misclassification score, or as part of hyperparameter tuning.
The cache naturally supports these types of queries when the depth is fixed and the size is increased, since the problem of computing a larger tree includes smaller trees as its subproblems, and all cache entries remain valid. In a sense, the framework incorporates incremental solving throughout its execution. However, not all entries can be kept once the global depth is increased.
When search and caching is performed, the results learned are (implicitly) only correct with respect to the maximum depth as the depth is not explicitly stored. Nevertheless, a portion of the cache entries remains valid when the depth is increased. Observe that for certain size values, increasing the depth is redundant and does not increase the search space. Intuitively, the limited tree size does not allow benefiting from a larger depth. Formally, given size and depth , if , then the set of all possible decision trees of size with depth is equivalent to the set of decision trees of size with depth , where . As a result, when incrementally computing a globally optimal decision tree of depth and the cache of the computation of a tree of depth is available with , we keep all cache entries such that , where is the corresponding path of the entry, and discard the rest.
4.7 Node Selection Strategy
Given a feature for a node and the size allocation for its children, the algorithm decides on which child node to recurse on first. Our search strategy is a variant of postorder traversal, labelled dynamic postorder, which dynamically decides which child node to visit first. The idea is to prioritise the child node that has the heuristicallydetermined largest potential to improve the current decision tree, which in turn leads to a higher chance to prune to the other sibling. The potential is computed as the gap between the upper and lower bound of the child. Note that the upper bound of the parent is used as the upper bound of its children. The lower bound for a child is retrieved from the cache. This provided consistent improvements in runtime (Section 5.2.3) when compared to the strategy used in DL8.5 (Aglin et al. (2020)), which always visits the left subtree before the right subtree.
4.8 Feature and Size Selection
For a given node, each possible tree configuration (a feature and the size of its children) is considered one at a time, unless the node is pruned or the optimal solution is retrieved from the cache (see Subsection 4.6). The order in which tree configurations are explored may have an impact on performance, as evidenced in most search algorithms in general.
In our experiments, the simplest variants performed the best: features are selected in the order as given in the dataset, and the size is distributed by considering increasingly larger right subtrees as described in Algorithm 1. Alternative options are possible, such as ordering the features according to their corresponding Gini coefficient as done in heuristic algorithms, but none of these alternatives lead to sizable benefits over the simplest strategies. This is discussed in more detail in the experimental section (Section 5.2.3).
4.9 Extensions
To ease the presentation, we discussed our framework in the context of binary classification on binary datasets. In this section, we discuss extensions to general settings.
4.9.1 General Data
The input to our framework is a dataset containing binary features. Datasets with continuous and/or categorical features are binarised. In our work, this done using a supervised discretisation algorithm based on the Minimum Description Length Principle (MDLP) by Fayyad and Irani (1993)
, effectively converting each feature into a categorical feature based on the statistical significance of the feature values for the class, and then using a onehot encoding to binarise the features. We use the MDLP implementation available in the R programming package (
Kim (2015)).Note that (univariate) decision tree algorithms implicitly binarise the dataset during learning. Each feature node is assigned a predicate that evaluates whether a given feature value meets a threshold, which can be seen as a binary feature. When binarising the dataset, the possible decisions are decided upfront rather than during execution.
4.9.2 MultiClassification
To extend the algorithm for multiclassification, the key step is to generalise Algorithm 3 to compute frequency counters for each class. Equations analogous to Equations 2—8 are devised to compute the misclassification scores. Since classes partition the data, the complexity results remain valid for multiclassification.
4.9.3 Additional Constraints and Objective Functions
Our framework may be modified to support constraints and objective functions that can be expressed in the dynamic programming formulation (Eq. 1) and solved using a (variant) of the specialised algorithm (Section 4.3).
Furthermore, objective functions that have a dependency on the number of nodes can also be handled. An alternate objective for optimal decision trees is
(18) 
which balances the size of the decision tree against the misclassifications. This objective was used in the original CART paper (Breiman et al. (1984)) and discussed in some of the other optimal decision tree works (Bertsimas and Dunn (2017); Hu et al. (2019)). While we could extend our solution to directly work with the sparse objective, we implemented a nonintrusive modification. We compute a series of decision trees minimising the misclassification score using nodes, where , and then evaluate each solution with respect to the objective and select the best. Our framework supports incremental solving, allowing cached subtrees from computing a decision tree with nodes to be reused when considering nodes. Another way of viewing the process is to consider it as a hyperparameter tuning procedure where the resulting trees are evaluated with respect to the new objective. A similar procedure was used by Bertsimas and Dunn (2017) to tune their integer programming approach.
We discussed multiclassification and the above objective function as special cases. Additional examples may include imposing a minimum number of instances per node, different linear penalties for misclassifying classes rather than treating each misclassification equally, and adding fairness objectives and constraints as in the work of Aghaei et al. (2019).
5 Computational Study
The goal of this section is to compare the performance of our method with the stateoftheart and empirically evaluate the scalability of our algorithm and the benefits of incremental computation and lower bounding. With this in mind, we designed three major themes to investigate, each addressing a unique set of questions: variations and scalability of our approach, effectiveness compared to the stateoftheart optimal classification tree algorithms, and outofsample accuracy as compared to heuristicallyobtained decision trees and random forests.
5.1 Datasets and Computational Environment
We use publicly available datasets used in previous works (Bertsimas and Dunn (2017); Verwer and Zhang (2019); Narodytska et al. (2018); Aglin et al. (2020); Hu et al. (2019)), most of which are available from the UCI and CP4IM repositories. The datasets include 85 classification problems with a mixture of binary, categorical, and continuous features. Datasets with categorical and/or continuous features are converted into binary datasets as discussed in Section 4.9.1. Datasets with missing values were excluded from experimentation. We note that some benchmarks appeared in previous works under different binarisation techniques or simplifications, e.g., multiclassification turned into binaryclassification using a ‘oneversusall’ scheme or a subset of the features were removed. For these cases, we include the different versions as separate datasets. The binarised datasets together with the binarisation script will be readily available soon (please contact the first author in the meantime).
Experiments were run on an Intel i77700HQ CPU with 32 GB of RAM running one algorithm at a time using one processor. The timeout was set to ten minutes except for the hyperparameter tuning where no limit was enforced. In the following, we dedicate a separate subsection to each of the three major experimental topics.
5.2 Variations of Our Algorithm and Scalability
The aim of this subsection is to investigate variations of our approach, determine the effectiveness of the introduced incremental computation and lower bounding techniques, analyse scalability, and impact of the feature and node selection strategies.
In the first part, we consider the algorithm without incremental or similaritybased lower bound computation, and observe the effect of adding these techniques to the runtime. In the second part, we discuss the impact of the number of instances, depth, and features with respect to the runtime. In the third part, we fix the algorithm parameters to their default settings and vary either the feature or node selection strategy.
We note that the default setting of our algorithm uses all techniques presented in the paper and the inorder feature selection and dynamic postorder node selection strategy.
5.2.1 Part One: Incremental and Lower Bound Computation
In Table 1, we show the effect of incrementally computing the frequency counters (Section 4.4.3) rather than doing it from scratch in each iteration, and combining it with our similaritybased lowerbound (Section 4.5).
The trend across all benchmarks is uniform, as each addition to the algorithm improves the runtime considerably. Incremental computation is useful as the splits considering two features might only differ in a small number of instances. Thus performing minor changes to the previously computed frequency counters is more favourable than recomputing from scratch. For similar reasons, the lower bound works well as it manages to identify cases where computing a subtree is unnecessary unless it deviates enough from the previously computed subtree.
An important observation that contributes to these positive points is that computing the difference of two sets of instances can be done in linear time with respect to the number of instances, which is comparatively inexpensive when considering the time spent on computing optimal subtrees of depth two using the specialised algorithm, but can save a significant amount of computation.
Name  OurBaseline  IncFQ  IncFQ+SimilarLB  

anneal  812  93  2  4  1  
audiology  216  148  2  7  1  1 
australiancredit  653  125  2  10  3  2 
breastwisconsin  683  120  2  8  2  1 
diabetes  768  112  2  11  4  3 
germancredit  1000  112  2  7  3  3 
heartcleveland  296  95  2  2  1  1 
hypothyroid  3247  88  2  12  3  2 
ionosphere  351  445  2  —  201  101 
krvskp  3196  73  2  7  2  1 
letter  20000  224  2  —  515  333 
mushroom  8124  119  2  2  1  1 
pendigits  7494  216  2  —  172  95 
splice1  3190  287  2  300  168  163 
vehicle  846  252  2  180  36  12 
yeast  1484  89  2  8  3  2 
biodeg  1055  81  2  2  1  1 
default_credit  30000  44  4  6  5  7 
HTRU_2  17898  57  2  2  2  3 
Ionosphere  351  143  2  4  2  2 
magic04  19020  79  2  6  4  6 
spambase  4601  132  2  27  11  11 
Statlog_satellite  4435  385  6  —  611  276 
Statlog_shuttle  43500  181  7  130  72  63 
appendicitisun  106  530  2  239  212  9 
australianun  690  1163  2  —  —  530 
backacheun  180  475  2  163  143  12 
cancerun  683  89  2  1  
cleveun  303  395  2  79  70  10 
colicun  368  415  2  111  91  22 
heartstatlogun  270  381  2  68  59  8 
hepatitisun  155  361  2  55  47  6 
hungarianun  294  330  2  39  34  6 
newthroidun  215  334  3  125  117  13 
promotersun  106  334  2  26  21  2 
shuttleMun  14500  691  2  —  —  169 
5.2.2 Part Two: Scalability
We investigate the sensitivity of our algorithm with respect to the number of instances and maximum depth. In Table 2, results are shown when our algorithm is run to compute trees of on datasets where instances are duplicated times. We note that trees with depth three are omitted as these are computed within seconds. The results indicate a linear dependency with the number of instances for the majority of the datasets. As most of the computational time is spent in repeatedly solving optimal subtrees of depth two (Section 4.4), the finding is consistent with the theoretical complexity (Proposition 4.4.2
). This is a notable improvement over generic optimisation approaches, such as integer programming or SAT. The latter may exhibit an exponential runtime dependency on the number of instances as new binary variables are introduced for each instance, and typically do not consider datasets with more than a thousand instances.
Note that the experiments regarding the scalability with respect to the number of instances are merely indicative. In practice, however, introducing more instances might implicitly increase or decrease the number of binary features in the discretisation and have an effect on shaping the structure of the dataset, both of which may impact positively or negatively the running time. The results do show that the bottleneck of the approach is not necessarily in the number of instances.
In contrast to the number of instances, the depth consistently has a large impact on the running time. The number of possible decision trees grows exponentially as the depth increases, which is reflected in the computational experiments. For example, our approach computes depththree trees within seconds, but the runtimes go up notably for depth four and five. In previous works, depth four has been seen as the benchmark value, but with MurTree such trees are computable within a reasonable time, and in many cases greater depths are also possible.
Apart from the depth, another important factor is the number of binary features, which additionally dictates the number of possible decision trees necessary to explore to find the optimal tree. As the ability of our techniques to prune and reduce computational time depends on the structure of the dataset, it is difficult to artificially increase the number of features and show the dependency. For example, duplicating features would not lead to conclusive statements on the impact of the number of features on runtime, as our lower bounding mechanism would trivially prune these features. We instead refer to the computational complexity of our algorithm from Proposition 4.4.2 and the number of possible decision trees as an indicative measure of the influence of the number of binary features and sparsity of the feature vectors on the runtime. Note that some of the instances contain a high number of features, e.g., australianun has 1163 binary features, but it is still within practical reach.
Overall, we conclude that our approach scales reasonably well for the tested datasets with depth four, having computed the optimal decision tree for the majority of the instances within seconds, and all within ten minutes. Deeper optimal decision trees of depth five or greater remain a challenge.
Name  Depth=4  Depth=5  
anneal  812  93  2  1  1  5  9  16  
australiancredit  653  125  2  2  3  6  63  92  162 
breastwisconsin  683  120  2  1  1  2  2  4  6 
diabetes  768  112  2  3  4  8  101  160  296 
germancredit  1000  112  2  3  5  9  111  187  331 
heartcleveland  296  95  2  1  1  1  9  13  20 
hypothyroid  3247  88  2  2  4  9  48  95  186 
ionosphere  351  445  2  101  163  279  215  336  623 
krvskp  3196  73  2  1  3  5  24  46  92 
letter  20000  224  2  333  —  —  —  —  — 
mushroom  8124  119  2  1  2  4  1  2  4 
pendigits  7494  216  2  95  166  314  415  —  — 
soybean  630  50  2  1  1  2  
splice1  3190  287  2  163  297  572  —  —  — 
tictactoe  958  27  2  1  1  2  
vehicle  846  252  2  12  21  44  446  —  — 
vote  435  48  2  1  1  2  
yeast  1484  89  2  2  4  6  77  136  226 
ficobinary  10459  17  2  2  4  8  
bank_conv  4521  26  2  1  1  2  3  7  
biodeg  1055  81  2  1  2  4  32  53  90 
default_credit  30000  44  4  7  19  98  114  322  — 
HTRU_2  17898  57  2  3  7 
Comments
There are no comments yet.