Masked Bayesian Neural Networks : Computation and Optimality

06/02/2022
by   Insung Kong, et al.
22

As data size and computing power increase, the architectures of deep neural networks (DNNs) have been getting more complex and huge, and thus there is a growing need to simplify such complex and huge DNNs. In this paper, we propose a novel sparse Bayesian neural network (BNN) which searches a good DNN with an appropriate complexity. We employ the masking variables at each node which can turn off some nodes according to the posterior distribution to yield a nodewise sparse DNN. We devise a prior distribution such that the posterior distribution has theoretical optimalities (i.e. minimax optimality and adaptiveness), and develop an efficient MCMC algorithm. By analyzing several benchmark datasets, we illustrate that the proposed BNN performs well compared to other existing methods in the sense that it discovers well condensed DNN architectures with similar prediction accuracy and uncertainty quantification compared to large DNNs.

READ FULL TEXT
research
05/24/2023

Masked Bayesian Neural Networks : Theoretical Guarantee and its Posterior Inference

Bayesian approaches for learning deep neural networks (BNN) have been re...
research
06/12/2022

Density Regression and Uncertainty Quantification with Bayesian Deep Noise Neural Networks

Deep neural network (DNN) models have achieved state-of-the-art predicti...
research
04/26/2022

PAC-Bayes training for neural networks: sparsity and uncertainty quantification

We study the Gibbs posterior distribution from PAC-Bayes theory for spar...
research
11/19/2020

Application of Deep Learning-based Interpolation Methods to Nearshore Bathymetry

Nearshore bathymetry, the topography of the ocean floor in coastal zones...
research
07/20/2021

A Bayesian Approach to Invariant Deep Neural Networks

We propose a novel Bayesian neural network architecture that can learn i...
research
06/07/2019

DropConnect Is Effective in Modeling Uncertainty of Bayesian Deep Networks

Deep neural networks (DNNs) have achieved state-of-the-art performances ...
research
04/13/2021

An acceleration strategy for randomize-then-optimize sampling via deep neural networks

Randomize-then-optimize (RTO) is widely used for sampling from posterior...

Please sign up or login with your details

Forgot password? Click here to reset