Conditional Variational Inference with Adaptive Truncation for Bayesian Nonparametric Models

01/13/2020
by   Jones Yirui Liu, et al.
0

The scalable inference for Bayesian nonparametric models with big data is still challenging. Current variational inference methods fail to characterise the correlation structure among latent variables due to the mean-field setting and cannot infer the true posterior dimension because of the universal truncation. To overcome these limitations, we build a general framework to infer Bayesian nonparametric models by maximising the proposed nonparametric evidence lower bound, and then develop a novel approach by combining Monte Carlo sampling and stochastic variational inference framework. Our method has several advantages over the traditional online variational inference method. First, it achieves a smaller divergence between variational distributions and the true posterior by factorising variational distributions under the conditional setting instead of the mean-field setting to capture the correlation pattern. Second, it reduces the risk of underfitting or overfitting by truncating the dimension adaptively rather than using a prespecified truncated dimension for all latent variables. Third, it reduces the computational complexity by approximating the posterior functionally instead of updating the stick-breaking parameters individually. We apply the proposed method on hierarchical Dirichlet process and gamma–Dirichlet process models, two essential Bayesian nonparametric models in topic analysis. The empirical study on three large datasets including arXiv, New York Times and Wikipedia reveals that our proposed method substantially outperforms its competitor in terms of lower perplexity and much clearer topic-words clustering.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/29/2012

Stochastic Variational Inference

We develop stochastic variational inference, a scalable algorithm for ap...
research
02/02/2018

An Instability in Variational Inference for Topic Models

Topic models are Bayesian models that are frequently used to capture the...
research
09/10/2018

Collapsed Variational Inference for Nonparametric Bayesian Group Factor Analysis

Group factor analysis (GFA) methods have been widely used to infer the c...
research
03/24/2011

The Discrete Infinite Logistic Normal Distribution

We present the discrete infinite logistic normal distribution (DILN), a ...
research
10/21/2015

Multiple co-clustering based on nonparametric mixture models with heterogeneous marginal distributions

We propose a novel method for multiple clustering that assumes a co-clus...
research
06/26/2015

An Empirical Study of Stochastic Variational Algorithms for the Beta Bernoulli Process

Stochastic variational inference (SVI) is emerging as the most promising...
research
09/12/2018

Discretely Relaxing Continuous Variables for tractable Variational Inference

We explore a new research direction in Bayesian variational inference wi...

Please sign up or login with your details

Forgot password? Click here to reset