Provable guarantees for decision tree induction: the agnostic setting

06/01/2020
by   Guy Blanc, et al.
0

We give strengthened provable guarantees on the performance of widely employed and empirically successful top-down decision tree learning heuristics. While prior works have focused on the realizable setting, we consider the more realistic and challenging agnostic setting. We show that for all monotone functions f and parameters s∈ℕ, these heuristics construct a decision tree of size s^Õ((log s)/ε^2) that achieves error ≤𝗈𝗉𝗍_s + ε, where 𝗈𝗉𝗍_s denotes the error of the optimal size-s decision tree for f. Previously, such a guarantee was not known to be achievable by any algorithm, even one that is not based on top-down heuristics. We complement our algorithmic guarantee with a near-matching s^Ω̃(log s) lower bound.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/18/2019

Top-down induction of decision trees: rigorous guarantees and inherent limitations

Consider the following heuristic for building a decision tree for a func...
research
10/16/2020

Universal guarantees for decision tree induction via a higher-order splitting criterion

We propose a simple extension of top-down decision tree learning heurist...
research
11/03/2020

Estimating decision tree learnability with polylogarithmic sample complexity

We show that top-down decision tree learning heuristics are amenable to ...
research
06/17/2022

Popular decision tree algorithms are provably noise tolerant

Using the framework of boosting, we prove that all impurity-based decisi...
research
07/02/2021

Decision tree heuristics can fail, even in the smoothed setting

Greedy decision tree learning heuristics are mainstays of machine learni...
research
06/03/2011

An Analysis of Reduced Error Pruning

Top-down induction of decision trees has been observed to suffer from th...

Please sign up or login with your details

Forgot password? Click here to reset