DeepAI AI Chat
Log In Sign Up

Deep Exponential Families

by   Rajesh Ranganath, et al.

We describe deep exponential families (DEFs), a class of latent variable models that are inspired by the hidden structures used in deep neural networks. DEFs capture a hierarchy of dependencies between latent variables, and are easily generalized to many settings through exponential families. We perform inference using recent "black box" variational inference techniques. We then evaluate various DEFs on text and combine multiple DEFs into a model for pairwise recommendation data. In an extensive study, we show that going beyond one layer improves predictions for DEFs. We demonstrate that DEFs find interesting exploratory structure in large data sets, and give better predictive performance than state-of-the-art models.


page 1

page 2

page 3

page 4


Convolutional Deep Exponential Families

We describe convolutional deep exponential families (CDEFs) in this pape...

Deep generative models of genetic variation capture mutation effects

The functions of proteins and RNAs are determined by a myriad of interac...

Hierarchical Variational Models

Black box variational inference allows researchers to easily prototype a...

Recurrent Latent Variable Networks for Session-Based Recommendation

In this work, we attempt to ameliorate the impact of data sparsity in th...

Clustering above Exponential Families with Tempered Exponential Measures

The link with exponential families has allowed k-means clustering to be ...

Overpruning in Variational Bayesian Neural Networks

The motivations for using variational inference (VI) in neural networks ...

A Forest Mixture Bound for Block-Free Parallel Inference

Coordinate ascent variational inference is an important algorithm for in...

Code Repositories


? Deep exponential family models in MXNet/Gluon ?

view repo