Streaming Bayesian inference: theoretical limits and mini-batch approximate message-passing

06/02/2017
by   Andre Manoel, et al.
0

In statistical learning for real-world large-scale data problems, one must often resort to "streaming" algorithms which operate sequentially on small batches of data. In this work, we present an analysis of the information-theoretic limits of mini-batch inference in the context of generalized linear models and low-rank matrix factorization. In a controlled Bayes-optimal setting, we characterize the optimal performance and phase transitions as a function of mini-batch size. We base part of our results on a detailed analysis of a mini-batch version of the approximate message-passing algorithm (Mini-AMP), which we introduce. Additionally, we show that this theoretical optimality carries over into real-data problems by illustrating that Mini-AMP is competitive with standard streaming algorithms for clustering.

READ FULL TEXT
research
03/13/2018

Dense Limit of the Dawid-Skene Model for Crowdsourcing and Regions of Sub-optimality of Message Passing Algorithms

Crowdsourcing is a strategy to categorize data through the contribution ...
research
03/02/2021

Mini-Batch Consistent Slot Set Encoder for Scalable Set Encoding

Most existing set encoding algorithms operate under the assumption that ...
research
05/10/2023

Phase transitions in the mini-batch size for sparse and dense neural networks

The use of mini-batches of data in training artificial neural networks i...
research
05/15/2018

On the glassy nature of the hard phase in inference problems

An algorithmically hard phase was described in a range of inference prob...
research
12/29/2017

A Unified Bayesian Inference Framework for Generalized Linear Models

In this letter, we present a unified Bayesian inference framework for ge...
research
10/03/2022

Bayes-optimal limits in structured PCA, and how to reach them

We study the paradigmatic spiked matrix model of principal components an...
research
06/22/2022

A view of mini-batch SGD via generating functions: conditions of convergence, phase transitions, benefit from negative momenta

Mini-batch SGD with momentum is a fundamental algorithm for learning lar...

Please sign up or login with your details

Forgot password? Click here to reset