Accelerated Linearized Laplace Approximation for Bayesian Deep Learning

10/23/2022
by   Zhijie Deng, et al.
0

Laplace approximation (LA) and its linearized variant (LLA) enable effortless adaptation of pretrained deep neural networks to Bayesian neural networks. The generalized Gauss-Newton (GGN) approximation is typically introduced to improve their tractability. However, LA and LLA are still confronted with non-trivial inefficiency issues and should rely on Kronecker-factored, diagonal, or even last-layer approximate GGN matrices in practical use. These approximations are likely to harm the fidelity of learning outcomes. To tackle this issue, inspired by the connections between LLA and neural tangent kernels (NTKs), we develop a Nystrom approximation to NTKs to accelerate LLA. Our method benefits from the capability of popular deep learning libraries for forward mode automatic differentiation, and enjoys reassuring theoretical guarantees. Extensive studies reflect the merits of the proposed method in aspects of both scalability and performance. Our method can even scale up to architectures like vision transformers. We also offer valuable ablation studies to diagnose our method. Code is available at <https://github.com/thudzj/ELLA>.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/24/2023

Variational Linearized Laplace Approximation for Bayesian Deep Learning

Pre-trained deep neural networks can be adapted to perform uncertainty e...
research
07/21/2020

Disentangling the Gauss-Newton Method and Approximate Inference for Neural Networks

In this thesis, we disentangle the generalized Gauss-Newton and approxim...
research
06/28/2021

Laplace Redux – Effortless Bayesian Deep Learning

Bayesian formulations of deep learning have been shown to have compellin...
research
05/20/2018

Online Structured Laplace Approximations For Overcoming Catastrophic Forgetting

We introduce the Kronecker factored online Laplace approximation for ove...
research
06/12/2023

Riemannian Laplace approximations for Bayesian neural networks

Bayesian neural networks often approximate the weight-posterior with a G...
research
06/24/2021

Lettuce: PyTorch-based Lattice Boltzmann Framework

The lattice Boltzmann method (LBM) is an efficient simulation technique ...
research
06/17/2022

Adapting the Linearised Laplace Model Evidence for Modern Deep Learning

The linearised Laplace method for estimating model uncertainty has recei...

Please sign up or login with your details

Forgot password? Click here to reset