Parallel Gaussian Process Regression for Big Data: Low-Rank Representation Meets Markov Approximation

11/17/2014
by   Kian Hsiang Low, et al.
0

The expressive power of a Gaussian process (GP) model comes at a cost of poor scalability in the data size. To improve its scalability, this paper presents a low-rank-cum-Markov approximation (LMA) of the GP model that is novel in leveraging the dual computational advantages stemming from complementing a low-rank approximate representation of the full-rank GP based on a support set of inputs with a Markov approximation of the resulting residual process; the latter approximation is guaranteed to be closest in the Kullback-Leibler distance criterion subject to some constraint and is considerably more refined than that of existing sparse GP models utilizing low-rank representations due to its more relaxed conditional independence assumption (especially with larger data). As a result, our LMA method can trade off between the size of the support set and the order of the Markov property to (a) incur lower computational cost than such sparse GP models while achieving predictive performance comparable to them and (b) accurately represent features/patterns of any scale. Interestingly, varying the Markov order produces a spectrum of LMAs with PIC approximation and full-rank GP at the two extremes. An advantage of our LMA method is that it is amenable to parallelization on multiple machines/cores, thereby gaining greater scalability. Empirical evaluation on three real-world datasets in clusters of up to 32 computing nodes shows that our centralized and parallel LMA methods are significantly more time-efficient and scalable than state-of-the-art sparse and full-rank GP regression methods while achieving comparable predictive performances.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/09/2014

Parallel Gaussian Process Regression with Low-Rank Covariance Matrix Approximations

Gaussian processes (GP) are Bayesian non-parametric models that are wide...
research
11/18/2016

A Generalized Stochastic Variational Bayesian Hyperparameter Learning Framework for Sparse Spectrum Gaussian Process Regression

While much research effort has been dedicated to scaling up sparse Gauss...
research
04/18/2012

EigenGP: Sparse Gaussian process models with data-dependent eigenfunctions

Gaussian processes (GPs) provide a nonparametric representation of funct...
research
11/03/2018

Understanding and Comparing Scalable Gaussian Process Regression for Big Data

As a non-parametric Bayesian model which produces informative predictive...
research
04/21/2014

GP-Localize: Persistent Mobile Robot Localization using Online Sparse Gaussian Process Observation Model

Central to robot exploration and mapping is the task of persistent local...
research
07/26/2022

Large-Scale Low-Rank Gaussian Process Prediction with Support Points

Low-rank approximation is a popular strategy to tackle the "big n proble...
research
08/09/2014

Decentralized Data Fusion and Active Sensing with Mobile Sensors for Modeling and Predicting Spatiotemporal Traffic Phenomena

The problem of modeling and predicting spatiotemporal traffic phenomena ...

Please sign up or login with your details

Forgot password? Click here to reset