Sparsifying Bayesian neural networks with latent binary variables and normalizing flows

05/05/2023
by   Lars Skaaret-Lund, et al.
0

Artificial neural networks (ANNs) are powerful machine learning methods used in many modern applications such as facial recognition, machine translation, and cancer diagnostics. A common issue with ANNs is that they usually have millions or billions of trainable parameters, and therefore tend to overfit to the training data. This is especially problematic in applications where it is important to have reliable uncertainty estimates. Bayesian neural networks (BNN) can improve on this, since they incorporate parameter uncertainty. In addition, latent binary Bayesian neural networks (LBBNN) also take into account structural uncertainty by allowing the weights to be turned on or off, enabling inference in the joint space of weights and structures. In this paper, we will consider two extensions to the LBBNN method: Firstly, by using the local reparametrization trick (LRT) to sample the hidden units directly, we get a more computationally efficient algorithm. More importantly, by using normalizing flows on the variational posterior distribution of the LBBNN parameters, the network learns a more flexible variational posterior distribution than the mean field Gaussian. Experimental results show that this improves predictive power compared to the LBBNN method, while also obtaining more sparse networks. We perform two simulation studies. In the first study, we consider variable selection in a logistic regression setting, where the more flexible variational distribution leads to improved results. In the second study, we compare predictive uncertainty based on data generated from two-dimensional Gaussian distributions. Here, we argue that our Bayesian methods lead to more realistic estimates of predictive uncertainty.

READ FULL TEXT

page 13

page 15

page 16

page 17

page 19

research
01/09/2023

Constraining cosmological parameters from N-body simulations with Variational Bayesian Neural Networks

Methods based on Deep Learning have recently been applied on astrophysic...
research
03/06/2017

Multiplicative Normalizing Flows for Variational Bayesian Neural Networks

We reinterpret multiplicative noise in neural networks as auxiliary rand...
research
03/02/2023

Variational EP with Probabilistic Backpropagation for Bayesian Neural Networks

I propose a novel approach for nonlinear Logistic regression using a two...
research
09/13/2019

Adversarial α-divergence Minimization for Bayesian Approximate Inference

Neural networks are popular models for regression. They are often traine...
research
10/12/2018

Predictive Uncertainty through Quantization

High-risk domains require reliable confidence estimates from predictive ...
research
03/04/2021

Approximate Bayesian Conditional Copulas

Copula models are flexible tools to represent complex structures of depe...
research
05/14/2020

Constraining the Reionization History using Bayesian Normalizing Flows

The next generation 21 cm surveys open a new window onto the early stage...

Please sign up or login with your details

Forgot password? Click here to reset