Compromise-free Bayesian neural networks

04/25/2020
by   Kamran Javid, et al.
0

We conduct a thorough analysis of the relationship between the out-of-sample performance and the Bayesian evidence (marginal likelihood) of Bayesian neural networks (BNNs), as well as looking at the performance of ensembles of BNNs, both using the Boston housing dataset. Using the state-of-the-art in nested sampling, we numerically sample the full (non-Gaussian and multimodal) network posterior and obtain numerical estimates of the Bayesian evidence, considering network models with up to 156 trainable parameters. The networks have between zero and four hidden layers, either tanh or ReLU activation functions, and with and without hierarchical priors. The ensembles of BNNs are obtained by determining the posterior distribution over networks, from the posterior samples of individual BNNs re-weighted by the associated Bayesian evidence values. There is good correlation between out-of-sample performance and evidence, as well as a remarkable symmetry between the evidence versus model size and out-of-sample performance versus model size planes. Networks with ReLU activation functions have consistently higher evidences than those with tanh functions, and this is reflected in their out-of-sample performance. Ensembling over architectures acts to further improve performance relative to the individual BNNs.

READ FULL TEXT

page 5

page 12

research
10/26/2021

Periodic Activation Functions Induce Stationarity

Neural network models are known to reinforce hidden data biases, making ...
research
10/14/2020

Effects of the Nonlinearity in Activation Functions on the Performance of Deep Learning Models

The nonlinearity of activation functions used in deep learning models ar...
research
03/29/2021

Comparison of different convolutional neural network activation functions and methods for building ensembles

Recently, much attention has been devoted to finding highly efficient an...
research
09/14/2020

Complexity Measures for Neural Networks with General Activation Functions Using Path-based Norms

A simple approach is proposed to obtain complexity controls for neural n...
research
03/15/2023

Bayesian Quadrature for Neural Ensemble Search

Ensembling can improve the performance of Neural Networks, but existing ...
research
12/29/2022

Bayesian Interpolation with Deep Linear Networks

This article concerns Bayesian inference using deep linear networks with...
research
02/18/2020

A Neural Network Based on First Principles

In this paper, a Neural network is derived from first principles, assumi...

Please sign up or login with your details

Forgot password? Click here to reset