Stationary Activations for Uncertainty Calibration in Deep Learning

10/19/2020
by   Lassi Meronen, et al.
0

We introduce a new family of non-linear neural network activation functions that mimic the properties induced by the widely-used Matérn family of kernels in Gaussian process (GP) models. This class spans a range of locally stationary models of various degrees of mean-square differentiability. We show an explicit link to the corresponding GP models in the case that the network consists of one infinitely wide hidden layer. In the limit of infinite smoothness the Matérn family results in the RBF kernel, and in this case we recover RBF activations. Matérn activation functions result in similar appealing properties to their counterparts in GP models, and we demonstrate that the local stationarity property together with limited mean-square differentiability shows both good performance and uncertainty calibration in Bayesian deep learning tasks. In particular, local stationarity helps calibrate out-of-distribution (OOD) uncertainty. We demonstrate these properties on classification and regression benchmarks and a radar emitter classification task.

READ FULL TEXT

page 2

page 8

page 19

page 24

page 25

research
09/26/2022

A connection between probability, physics and neural networks

We illustrate an approach that can be exploited for constructing neural ...
research
12/01/2021

Mixed neural network Gaussian processes

This paper makes two contributions. Firstly, it introduces mixed composi...
research
07/26/2020

Regularized Flexible Activation Function Combinations for Deep Neural Networks

Activation in deep neural networks is fundamental to achieving non-linea...
research
05/24/2022

Imposing Gaussian Pre-Activations in a Neural Network

The goal of the present work is to propose a way to modify both the init...
research
09/09/2022

Fast Neural Kernel Embeddings for General Activations

Infinite width limit has shed light on generalization and optimization a...
research
02/22/2021

Elementary superexpressive activations

We call a finite family of activation functions superexpressive if any m...
research
12/09/2020

Know Your Limits: Monotonicity Softmax Make Neural Classifiers Overconfident on OOD Data

A crucial requirement for reliable deployment of deep learning models fo...

Please sign up or login with your details

Forgot password? Click here to reset