Evidential Softmax for Sparse Multimodal Distributions in Deep Generative Models

10/27/2021
by   Phil Chen, et al.
0

Many applications of generative models rely on the marginalization of their high-dimensional output probability distributions. Normalization functions that yield sparse probability distributions can make exact marginalization more computationally tractable. However, sparse normalization functions usually require alternative loss functions for training since the log-likelihood is undefined for sparse probability distributions. Furthermore, many sparse normalization functions often collapse the multimodality of distributions. In this work, we present ev-softmax, a sparse normalization function that preserves the multimodality of probability distributions. We derive its properties, including its gradient in closed-form, and introduce a continuous family of approximations to ev-softmax that have full support and can be trained with probabilistic loss functions such as negative log-likelihood and Kullback-Leibler divergence. We evaluate our method on a variety of generative models, including variational autoencoders and auto-regressive architectures. Our method outperforms existing dense and sparse normalization techniques in distributional accuracy. We demonstrate that ev-softmax successfully reduces the dimensionality of probability distributions while maintaining multimodality.

READ FULL TEXT
research
01/28/2021

Probabilistic Data with Continuous Distributions

Statistical models of real world data typically involve continuous proba...
research
11/16/2015

An Exploration of Softmax Alternatives Belonging to the Spherical Loss Family

In a multi-class classification problem, it is standard to model the out...
research
02/13/2023

Variational Mixture of HyperGenerators for Learning Distributions Over Functions

Recent approaches build on implicit neural representations (INRs) to pro...
research
10/29/2021

Resampling Base Distributions of Normalizing Flows

Normalizing flows are a popular class of models for approximating probab...
research
06/17/2020

Probabilistic orientation estimation with matrix Fisher distributions

This paper focuses on estimating probability distributions over the set ...
research
05/19/2017

The Kernel Mixture Network: A Nonparametric Method for Conditional Density Estimation of Continuous Random Variables

This paper introduces the kernel mixture network, a new method for nonpa...
research
06/26/2016

Exact gradient updates in time independent of output size for the spherical loss family

An important class of problems involves training deep neural networks wi...

Please sign up or login with your details

Forgot password? Click here to reset