Interpretable Reinforcement Learning via Differentiable Decision Trees

Decision trees are ubiquitous in machine learning for their ease of use and interpretability; however, they are not typically implemented in reinforcement learning because they cannot be updated via stochastic gradient descent. Traditional applications of decision trees for reinforcement learning have focused instead on making commitments to decision boundaries as the tree is grown one layer at a time. We overcome this critical limitation by allowing for a gradient update over the entire tree structure that improves sample complexity when a tree is fuzzy and interpretability when sharp. We offer three key contributions towards this goal. First, we motivate the need for policy gradient-based learning by examining the theoretical properties of gradient descent over differentiable decision trees. Second, we introduce a regularization framework that yields interpretability via sparsity in the tree structure. Third, we demonstrate the ability to construct a decision tree via policy gradient in canonical reinforcement learning domains and supervised learning benchmarks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/19/2018

Deep Neural Decision Trees

Deep neural networks have been proven powerful at processing perceptual ...
research
01/23/2019

Stochastic Gradient Trees

We present an online algorithm that induces decision trees using gradien...
research
02/15/2021

Learning Accurate Decision Trees with Bandit Feedback via Quantized Gradient Descent

Decision trees provide a rich family of highly non-linear but efficient ...
research
05/05/2023

Learning Decision Trees with Gradient Descent

Decision Trees (DTs) are commonly used for many machine learning tasks d...
research
12/14/2020

Evolutionary learning of interpretable decision trees

Reinforcement learning techniques achieved human-level performance in se...
research
09/26/2014

Autoencoder Trees

We discuss an autoencoder model in which the encoding and decoding funct...
research
06/09/2022

There is no Accuracy-Interpretability Tradeoff in Reinforcement Learning for Mazes

Interpretability is an essential building block for trustworthiness in r...

Please sign up or login with your details

Forgot password? Click here to reset