The Tree Ensemble Layer: Differentiability meets Conditional Computation

02/18/2020
by   Hussein Hazimeh, et al.
1

Neural networks and tree ensembles are state-of-the-art learners, each with its unique statistical and computational advantages. We aim to combine these advantages by introducing a new layer for neural networks, composed of an ensemble of differentiable decision trees (a.k.a. soft trees). While differentiable trees demonstrate promising results in the literature, in practice they are typically slow in training and inference as they do not support conditional computation. We mitigate this issue by introducing a new sparse activation function for sample routing, and implement true conditional computation by developing specialized forward and backward propagation algorithms that exploit sparsity. Our efficient algorithms pave the way for jointly training over deep and wide tree ensembles using first-order methods (e.g., SGD). Experiments on 23 classification datasets indicate over 10x speed-ups compared to the differentiable trees used in the literature and over 20x reduction in the number of parameters compared to gradient boosted trees, while maintaining competitive performance. Moreover, experiments on CIFAR, MNIST, and Fashion MNIST indicate that replacing dense layers in CNNs with our tree layer reduces the test loss by 7-53 We provide an open-source TensorFlow implementation with a Keras API.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/19/2022

Flexible Modeling and Multitask Learning using Differentiable Tree Ensembles

Decision tree ensembles are widely used and competitive learning models....
research
09/10/2021

A Neural Tangent Kernel Perspective of Infinite Tree Ensembles

In practical situations, the ensemble tree model is one of the most popu...
research
07/25/2018

Conditional Information Gain Networks

Deep neural network models owe their representational power to the high ...
research
02/21/2023

Variational Boosted Soft Trees

Gradient boosting machines (GBMs) based on decision trees consistently d...
research
01/23/2019

Stochastic Gradient Trees

We present an online algorithm that induces decision trees using gradien...
research
07/17/2018

Adaptive Neural Trees

Deep neural networks and decision trees operate on largely separate para...
research
06/28/2014

Exponentially Increasing the Capacity-to-Computation Ratio for Conditional Computation in Deep Learning

Many state-of-the-art results obtained with deep networks are achieved w...

Please sign up or login with your details

Forgot password? Click here to reset