Adversarial Distillation of Bayesian Neural Network Posteriors

06/27/2018
by   Kuan-Chieh Wang, et al.
4

Bayesian neural networks (BNNs) allow us to reason about uncertainty in a principled way. Stochastic Gradient Langevin Dynamics (SGLD) enables efficient BNN learning by drawing samples from the BNN posterior using mini-batches. However, SGLD and its extensions require storage of many copies of the model parameters, a potentially prohibitive cost, especially for large neural networks. We propose a framework, Adversarial Posterior Distillation, to distill the SGLD samples using a Generative Adversarial Network (GAN). At test-time, samples are generated by the GAN. We show that this distillation framework incurs no loss in performance on recent BNN applications including anomaly detection, active learning, and defense against adversarial attacks. By construction, our framework not only distills the Bayesian predictive distribution, but the posterior itself. This allows one to compute quantities such as the approximate model variance, which is useful in downstream tasks. To our knowledge, these are the first results applying MCMC-based BNNs to the aforementioned downstream applications.

READ FULL TEXT
research
10/13/2017

Bayesian Hypernetworks

We propose Bayesian hypernetworks: a framework for approximate Bayesian ...
research
05/16/2020

Generalized Bayesian Posterior Expectation Distillation for Deep Neural Networks

In this paper, we present a general framework for distilling expectation...
research
01/27/2021

Evolutionary Generative Adversarial Networks with Crossover Based Knowledge Distillation

Generative Adversarial Networks (GAN) is an adversarial model, and it ha...
research
06/14/2015

Bayesian Dark Knowledge

We consider the problem of Bayesian parameter estimation for deep neural...
research
09/10/2023

DAD++: Improved Data-free Test Time Adversarial Defense

With the increasing deployment of deep neural networks in safety-critica...
research
08/25/2022

Adversarial Bayesian Simulation

In the absence of explicit or tractable likelihoods, Bayesians often res...
research
04/08/2023

PVD-AL: Progressive Volume Distillation with Active Learning for Efficient Conversion Between Different NeRF Architectures

Neural Radiance Fields (NeRF) have been widely adopted as practical and ...

Please sign up or login with your details

Forgot password? Click here to reset