DeepAI AI Chat
Log In Sign Up

Stein Point Markov Chain Monte Carlo

by   Wilson Ye Chen, et al.

An important task in machine learning and statistics is the approximation of a probability measure by an empirical measure supported on a discrete point set. Stein Points are a class of algorithms for this task, which proceed by sequentially minimising a Stein discrepancy between the empirical measure and the target and, hence, require the solution of a non-convex optimisation problem to obtain each new point. This paper removes the need to solve this optimisation problem by, instead, selecting each new point based on a Markov chain sample path. This significantly reduces the computational cost of Stein Points and leads to a suite of algorithms that are straightforward to implement. The new algorithms are illustrated on a set of challenging Bayesian inference problems, and rigorous theoretical guarantees of consistency are established.


Optimal quantisation of probability measures using maximum mean discrepancy

Several researchers have proposed minimisation of maximum mean discrepan...

Stein Points

An important task in computational statistics and machine learning is to...

Batch Selection for Parallelisation of Bayesian Quadrature

Integration over non-negative integrands is a central problem in machine...

Optimal Thinning of MCMC Output

The use of heuristics to assess the convergence and compress the output ...

Bayesian spatial voting model to characterize the legislative behavior of the Colombian Senate 2010-2014

This paper applies Bayesian methodologies to characterize the legislativ...

Probabilistic Inference on Noisy Time Series (PINTS)

Time series models are ubiquitous in science, arising in any situation w...