Superpolynomial Lower Bounds for Decision Tree Learning and Testing

10/12/2022
by   Caleb Koch, et al.
0

We establish new hardness results for decision tree optimization problems, adding to a line of work that dates back to Hyafil and Rivest in 1976. We prove, under randomized ETH, superpolynomial lower bounds for two basic problems: given an explicit representation of a function f and a generator for a distribution 𝒟, construct a small decision tree approximator for f under 𝒟, and decide if there is a small decision tree approximator for f under 𝒟. Our results imply new lower bounds for distribution-free PAC learning and testing of decision trees, settings in which the algorithm only has restricted access to f and 𝒟. Specifically, we show: n-variable size-s decision trees cannot be properly PAC learned in time n^Õ(loglog s), and depth-d decision trees cannot be tested in time exp(d^ O(1)). For learning, the previous best lower bound only ruled out poly(n)-time algorithms (Alekhnovich, Braverman, Feldman, Klivans, and Pitassi, 2009). For testing, recent work gives similar though incomparable bounds in the setting where f is random and 𝒟 is nonexplicit (Blais, Ferreira Pinto Jr., and Harms, 2021). Assuming a plausible conjecture on the hardness of Set-Cover, we show our lower bound for learning decision trees can be improved to n^Ω(log s), matching the best known upper bound of n^O(log s) due to Ehrenfeucht and Haussler (1989). We obtain our results within a unified framework that leverages recent progress in two lines of work: the inapproximability of Set-Cover and XOR lemmas for query complexity. Our framework is versatile and yields results for related concept classes such as juntas and DNF formulas.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/20/2023

Superpolynomial Lower Bounds for Learning Monotone Classes

Koch, Strassle, and Tan [SODA 2023], show that, under the randomized exp...
research
10/18/2021

A cautionary tale on fitting decision trees to data from additive models: generalization lower bounds

Decision trees are important both as interpretable models amenable to hi...
research
07/09/2023

Properly Learning Decision Trees with Queries Is NP-Hard

We prove that it is NP-hard to properly PAC learn decision trees with qu...
research
02/14/2023

Bounds on Depth of Decision Trees Derived from Decision Rule Systems

Systems of decision rules and decision trees are widely used as a means ...
research
05/19/2018

Adaptively Pruning Features for Boosted Decision Trees

Boosted decision trees enjoy popularity in a variety of applications; ho...
research
11/22/2022

Query Complexity of Inversion Minimization on Trees

We consider the following computational problem: Given a rooted tree and...
research
03/27/2023

Lifting uniform learners via distributional decomposition

We show how any PAC learning algorithm that works under the uniform dist...

Please sign up or login with your details

Forgot password? Click here to reset