Stochastic geometry to generalize the Mondrian Process

02/03/2020
by   Eliza O'Reilly, et al.
0

The Mondrian process is a stochastic process that produces a recursive partition of space with random axis-aligned cuts. Random forests and Laplace kernel approximations built from the Mondrian process have led to efficient online learning methods and Bayesian optimization. By viewing the Mondrian process as a special case of the stable under iterated tessellation (STIT) process, we utilize tools from stochastic geometry to resolve three fundamental questions concern generalizability of the Mondrian process in machine learning. First, we show that the Mondrian process with general cut directions can be efficiently simulated, but it is unlikely to give rise to better classification or regression algorithms. Second, we characterize all possible kernels that generalizations of the Mondrian process can approximate. This includes, for instance, various forms of the weighted Laplace kernel and the exponential kernel. Third, we give an explicit formula for the density estimator arising from a Mondrian forest. This allows for precise comparisons between the Mondrian forest, the Mondrian kernel and the Laplace kernel in density estimation. Our paper calls for further developments at the novel intersection of stochastic geometry and machine learning.

READ FULL TEXT
research
06/16/2016

The Mondrian Kernel

We introduce the Mondrian kernel, a fast random feature approximation to...
research
10/11/2017

Improved Coresets for Kernel Density Estimates

We study the construction of coresets for kernel density estimates. That...
research
06/13/2019

Random Tessellation Forests

Space partitioning methods such as random forests and the Mondrian proce...
research
09/22/2020

Deep Neural Tangent Kernel and Laplace Kernel Have the Same RKHS

We prove that the reproducing kernel Hilbert spaces (RKHS) of a deep neu...
research
11/02/2021

Geometry-aware Bayesian Optimization in Robotics using Riemannian Matérn Kernels

Bayesian optimization is a data-efficient technique which can be used fo...
research
03/28/2014

Characteristic Kernels and Infinitely Divisible Distributions

We connect shift-invariant characteristic kernels to infinitely divisibl...
research
11/30/2018

Decision Forests Induce Characteristic Kernels

Decision forests are popular tools for classification and regression. Th...

Please sign up or login with your details

Forgot password? Click here to reset