Parallelising MCMC via Random Forests

11/21/2019
by   Wu Changye, et al.
0

For Bayesian computation in big data contexts, the divide-and-conquer MCMC concept splits the whole data set into batches, runs MCMC algorithms separately over each batch to produce samples of parameters, and combines them to produce an approximation of the target distribution. In this article, we embed random forests into this framework and use each subposterior/partial-posterior as a proposal distribution to implement importance sampling. Unlike the existing divide-and-conquer MCMC, our methods are based on scaled subposteriors, whose scale factors are not necessarily restricted to being equal to one or to the number of subsets. Through several experiments, we show that our methods work well with models ranging from Gaussian cases to strongly non-Gaussian cases, and include model misspecification.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/26/2018

ABC Samplers

This Chapter, "ABC Samplers", is to appear in the forthcoming Handbook o...
research
01/15/2018

Divide and Recombine for Large and Complex Data: Model Likelihood Functions using MCMC

In Divide & Recombine (D&R), big data are divided into subsets, each ana...
research
06/10/2015

Parallelizing MCMC with Random Partition Trees

The modern scale of data has brought new challenges to Bayesian inferenc...
research
08/31/2021

Bayesian learning of forest and tree graphical models

In Bayesian learning of Gaussian graphical model structure, it is common...
research
03/11/2019

Embarrassingly parallel MCMC using deep invertible transformations

While MCMC methods have become a main work-horse for Bayesian inference,...
research
07/22/2021

On the Scalability of Informed Importance Tempering

Informed MCMC methods have been proposed as scalable solutions to Bayesi...

Please sign up or login with your details

Forgot password? Click here to reset