Sparse Probabilistic Circuits via Pruning and Growing

11/22/2022
by   Meihua Dang, et al.
0

Probabilistic circuits (PCs) are a tractable representation of probability distributions allowing for exact and efficient computation of likelihoods and marginals. There has been significant recent progress on improving the scale and expressiveness of PCs. However, PC training performance plateaus as model size increases. We discover that most capacity in existing large PC structures is wasted: fully-connected parameter layers are only sparsely used. We propose two operations: pruning and growing, that exploit the sparsity of PC structures. Specifically, the pruning operation removes unimportant sub-networks of the PC for model compression and comes with theoretical guarantees. The growing operation increases model capacity by increasing the size of the latent space. By alternatingly applying pruning and growing, we increase the capacity that is meaningfully used, allowing us to significantly scale up PC learning. Empirically, our learner achieves state-of-the-art likelihoods on MNIST-family image datasets and on Penn Tree Bank language data compared to other PC learners and less tractable deep generative models such as flow-based models and variational autoencoders (VAEs).

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/10/2022

Scaling Up Probabilistic Circuits by Latent Variable Distillation

Probabilistic Circuits (PCs) are a unified framework for tractable proba...
research
02/16/2023

Understanding the Distillation Process from Deep Generative Models to Tractable Probabilistic Circuits

Probabilistic Circuits (PCs) are a general and unified computational fra...
research
04/13/2020

Einsum Networks: Fast and Scalable Learning of Tractable Probabilistic Circuits

Probabilistic circuits (PCs) are a promising avenue for probabilistic mo...
research
02/20/2020

Regularized Autoencoders via Relaxed Injective Probability Flow

Invertible flow-based generative models are an effective method for lear...
research
05/04/2021

Alternate Model Growth and Pruning for Efficient Training of Recommendation Systems

Deep learning recommendation systems at scale have provided remarkable g...
research
06/26/2020

On the Relationship Between Probabilistic Circuits and Determinantal Point Processes

Scaling probabilistic models to large realistic problems and datasets is...
research
09/28/2022

ButterflyFlow: Building Invertible Layers with Butterfly Matrices

Normalizing flows model complex probability distributions using maps obt...

Please sign up or login with your details

Forgot password? Click here to reset