Latent Variable Modeling with Diversity-Inducing Mutual Angular Regularization

12/23/2015
by   Pengtao Xie, et al.
0

Latent Variable Models (LVMs) are a large family of machine learning models providing a principled and effective way to extract underlying patterns, structure and knowledge from observed data. Due to the dramatic growth of volume and complexity of data, several new challenges have emerged and cannot be effectively addressed by existing LVMs: (1) How to capture long-tail patterns that carry crucial information when the popularity of patterns is distributed in a power-law fashion? (2) How to reduce model complexity and computational cost without compromising the modeling power of LVMs? (3) How to improve the interpretability and reduce the redundancy of discovered patterns? To addresses the three challenges discussed above, we develop a novel regularization technique for LVMs, which controls the geometry of the latent space during learning to enable the learned latent components of LVMs to be diverse in the sense that they are favored to be mutually different from each other, to accomplish long-tail coverage, low redundancy, and better interpretability. We propose a mutual angular regularizer (MAR) to encourage the components in LVMs to have larger mutual angles. The MAR is non-convex and non-smooth, entailing great challenges for optimization. To cope with this issue, we derive a smooth lower bound of the MAR and optimize the lower bound instead. We show that the monotonicity of the lower bound is closely aligned with the MAR to qualify the lower bound as a desirable surrogate of the MAR. Using neural network (NN) as an instance, we analyze how the MAR affects the generalization performance of NN. On two popular latent variable models --- restricted Boltzmann machine and distance metric learning, we demonstrate that MAR can effectively capture long-tail patterns, reduce model complexity without sacrificing expressivity and improve interpretability.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/13/2018

GILBO: One Metric to Measure Them All

We propose a simple, tractable lower bound on the mutual information con...
research
11/23/2017

Diversity-Promoting Bayesian Learning of Latent Variable Models

To address three important issues involved in latent variable models (LV...
research
04/12/2019

Information Theoretic Lower Bounds on Negative Log Likelihood

In this article we use rate-distortion theory, a branch of information t...
research
02/17/2020

Augmented Normalizing Flows: Bridging the Gap Between Generative Flows and Latent Variable Models

In this work, we propose a new family of generative flows on an augmente...
research
07/04/2023

Minimax rates for latent position estimation in the generalized random dot product graph

Latent space models play an important role in the modeling and analysis ...
research
08/10/2018

Multi-Channel Stochastic Variational Inference for the Joint Analysis of Heterogeneous Biomedical Data in Alzheimer's Disease

The joint analysis of biomedical data in Alzheimer's Disease (AD) is imp...
research
10/05/2020

Tight bounds on the mutual coherence of sensing matrices for Wigner D-functions on regular grids

Many practical sampling patterns for function approximation on the rotat...

Please sign up or login with your details

Forgot password? Click here to reset