On the Delta Method for Uncertainty Approximation in Deep Learning

12/02/2019
by   Geir K. Nilsen, et al.
0

The Delta method is a well known procedure used to quantify uncertainty in statistical models. The method has previously been applied in the context of neural networks, but has not reached much popularity in deep learning because of the sheer size of the Hessian matrix. In this paper, we propose a low cost variant of the method based on an approximate eigendecomposition of the positive curvature subspace of the Hessian matrix. The method has a computational complexity of O(KPN) time and O(KP) space, where K is the number of utilized Hessian eigenpairs, P is the number of model parameters and N is the number of training examples. Given that the model is L_2-regularized with rate λ/2, we provide a bound on the uncertainty approximation error given K. We show that when the smallest Hessian eigenvalue in the positive K/2-tail of the full spectrum, and the largest Hessian eigenvalue in the negative K/2-tail of the full spectrum are both approximately equal to λ, the error will be close to zero even when K≪ P . We demonstrate the method by a TensorFlow implementation, and show that meaningful rankings of images based on prediction uncertainty can be obtained for a convolutional neural network based MNIST classifier. We also observe that false positives have higher prediction uncertainty than true positives. This suggests that there is supplementing information in the uncertainty measure not captured by the probability alone.

READ FULL TEXT
research
01/12/2021

Using uncertainty estimation to reduce false positives in liver lesion detection

Despite the successes of deep learning techniques at detecting objects i...
research
05/14/2019

Efficient Computation of Hessian Matrices in TensorFlow

The Hessian matrix has a number of important applications in a variety o...
research
01/14/2020

On the Convex Behavior of Deep Neural Networks in Relation to the Layers' Width

The Hessian of neural networks can be decomposed into a sum of two matri...
research
12/14/2020

A spectral characterization and an approximation scheme for the Hessian eigenvalue

We revisit the k-Hessian eigenvalue problem on a smooth, bounded, (k-1)-...
research
01/29/2019

An Investigation into Neural Net Optimization via Hessian Eigenvalue Density

To understand the dynamics of optimization in deep neural networks, we d...
research
05/29/2023

SANE: The phases of gradient descent through Sharpness Adjusted Number of Effective parameters

Modern neural networks are undeniably successful. Numerous studies have ...
research
01/24/2019

Measurements of Three-Level Hierarchical Structure in the Outliers in the Spectrum of Deepnet Hessians

We consider deep classifying neural networks. We expose a structure in t...

Please sign up or login with your details

Forgot password? Click here to reset