A Jensen-Shannon Divergence Based Loss Function for Bayesian Neural Networks

09/23/2022
by   Ponkrshnan Thiagarajan, et al.
0

Kullback-Leibler (KL) divergence is widely used for variational inference of Bayesian Neural Networks (BNNs). However, the KL divergence has limitations such as unboundedness and asymmetry. We examine the Jensen-Shannon (JS) divergence that is more general, bounded, and symmetric. We formulate a novel loss function for BNNs based on the geometric JS divergence and show that the conventional KL divergence-based loss function is its special case. We evaluate the divergence part of the proposed loss function in a closed form for a Gaussian prior. For any other general prior, Monte Carlo approximations can be used. We provide algorithms for implementing both of these cases. We demonstrate that the proposed loss function offers an additional parameter that can be tuned to control the degree of regularisation. We derive the conditions under which the proposed loss function regularises better than the KL divergence-based loss function for Gaussian priors and posteriors. We demonstrate performance improvements over the state-of-the-art KL divergence-based BNN on the classification of a noisy CIFAR data set and a biased histopathology data set.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/10/2018

On the Universality of the Logistic Loss Function

A loss function measures the discrepancy between the true values (observ...
research
11/09/2022

A Diffeomorphic Flow-based Variational Framework for Multi-speaker Emotion Conversion

This paper introduces a new framework for non-parallel emotion conversio...
research
09/22/2021

On Reparameterization Invariant Bayesian Point Estimates and Credible Regions

This paper considers reparameterization invariant Bayesian point estimat...
research
10/14/2018

Bregman Divergence Bounds and the Universality of the Logarithmic Loss

A loss function measures the discrepancy between the true values and the...
research
05/01/2017

Nonlinear Kalman Filtering with Divergence Minimization

We consider the nonlinear Kalman filtering problem using Kullback-Leible...
research
08/29/2018

Centroid estimation based on symmetric KL divergence for Multinomial text classification problem

We define a new method to estimate centroid for text classification base...
research
07/09/2011

Loss-sensitive Training of Probabilistic Conditional Random Fields

We consider the problem of training probabilistic conditional random fie...

Please sign up or login with your details

Forgot password? Click here to reset