ALBU: An approximate Loopy Belief message passing algorithm for LDA to improve performance on small data sets

10/01/2021
by   Rebecca M. C. Taylor, et al.
0

Variational Bayes (VB) applied to latent Dirichlet allocation (LDA) has become the most popular algorithm for aspect modeling. While sufficiently successful in text topic extraction from large corpora, VB is less successful in identifying aspects in the presence of limited data. We present a novel variational message passing algorithm as applied to Latent Dirichlet Allocation (LDA) and compare it with the gold standard VB and collapsed Gibbs sampling. In situations where marginalisation leads to non-conjugate messages, we use ideas from sampling to derive approximate update equations. In cases where conjugacy holds, Loopy Belief update (LBU) (also known as Lauritzen-Spiegelhalter) is used. Our algorithm, ALBU (approximate LBU), has strong similarities with Variational Message Passing (VMP) (which is the message passing variant of VB). To compare the performance of the algorithms in the presence of limited data, we use data sets consisting of tweets and news groups. Additionally, to perform more fine grained evaluations and comparisons, we use simulations that enable comparisons with the ground truth via Kullback-Leibler divergence (KLD). Using coherence measures for the text corpora and KLD with the simulations we show that ALBU learns latent distributions more accurately than does VB, especially for smaller data sets.

READ FULL TEXT
research
08/19/2022

SimLDA: A tool for topic model evaluation

Variational Bayes (VB) applied to latent Dirichlet allocation (LDA) has ...
research
11/02/2021

Variational message passing (VMP) applied to LDA

Variational Bayes (VB) applied to latent Dirichlet allocation (LDA) is t...
research
04/01/2021

Bayesian Functional Principal Components Analysis via Variational Message Passing

Functional principal components analysis is a popular tool for inference...
research
06/27/2012

Rethinking Collapsed Variational Bayes Inference for LDA

We propose a novel interpretation of the collapsed variational Bayes inf...
research
07/27/2018

Pull Message Passing for Nonparametric Belief Propagation

We present a "pull" approach to approximate products of Gaussian mixture...
research
02/18/2022

A new LDA formulation with covariates

The Latent Dirichlet Allocation (LDA) model is a popular method for crea...
research
04/12/2020

Active Sampling for Pairwise Comparisons via Approximate Message Passing and Information Gain Maximization

Pairwise comparison data arise in many domains with subjective assessmen...

Please sign up or login with your details

Forgot password? Click here to reset