Large-Scale Distributed Bayesian Matrix Factorization using Stochastic Gradient MCMC

03/05/2015
by   Sungjin Ahn, et al.
0

Despite having various attractive qualities such as high prediction accuracy and the ability to quantify uncertainty and avoid over-fitting, Bayesian Matrix Factorization has not been widely adopted because of the prohibitive cost of inference. In this paper, we propose a scalable distributed Bayesian matrix factorization algorithm using stochastic gradient MCMC. Our algorithm, based on Distributed Stochastic Gradient Langevin Dynamics, can not only match the prediction accuracy of standard MCMC methods like Gibbs sampling, but at the same time is as fast and simple as stochastic gradient descent. In our experiments, we show that our algorithm can achieve the same level of prediction accuracy as Gibbs sampling an order of magnitude faster. We also show that our method reduces the prediction error as fast as distributed stochastic gradient descent, achieving a 4.1 Netflix dataset and an 1.8

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/06/2020

A High-Performance Implementation of Bayesian Matrix Factorization with Limited Communication

Matrix factorization is a very common machine learning technique in reco...
research
03/07/2019

On Transformations in Stochastic Gradient MCMC

Stochastic gradient Langevin dynamics (SGLD) is a widely used sampler fo...
research
02/13/2020

Fast Convergence for Langevin Diffusion with Matrix Manifold Structure

In this paper, we study the problem of sampling from distributions of th...
research
11/12/2013

DinTucker: Scaling up Gaussian process models on multidimensional arrays with billions of elements

Infinite Tucker Decomposition (InfTucker) and random function prior mode...
research
07/26/2018

A Collaborative Approach to Angel and Venture Capital Investment Recommendations

Matrix factorization was used to generate investment recommendations for...
research
09/05/2015

HAMSI: A Parallel Incremental Optimization Algorithm Using Quadratic Approximations for Solving Partially Separable Problems

We propose HAMSI (Hessian Approximated Multiple Subsets Iteration), whic...
research
12/13/2018

Stochastic Gradient Descent for Spectral Embedding with Implicit Orthogonality Constraint

In this paper, we propose a scalable algorithm for spectral embedding. T...

Please sign up or login with your details

Forgot password? Click here to reset