Bayesian Inference with Certifiable Adversarial Robustness

02/10/2021
by   Matthew Wicker, et al.
7

We consider adversarial training of deep neural networks through the lens of Bayesian learning, and present a principled framework for adversarial training of Bayesian Neural Networks (BNNs) with certifiable guarantees. We rely on techniques from constraint relaxation of non-convex optimisation problems and modify the standard cross-entropy error model to enforce posterior robustness to worst-case perturbations in ϵ-balls around input points. We illustrate how the resulting framework can be combined with methods commonly employed for approximate inference of BNNs. In an empirical investigation, we demonstrate that the presented approach enables training of certifiably robust models on MNIST, FashionMNIST and CIFAR-10 and can also be beneficial for uncertainty calibration. Our method is the first to directly train certifiable BNNs, thus facilitating their deployment in safety-critical applications.

READ FULL TEXT

page 7

page 14

page 15

page 17

research
04/20/2018

Learning More Robust Features with Adversarial Training

In recent years, it has been found that neural networks can be easily fo...
research
06/23/2023

Adversarial Robustness Certification for Bayesian Neural Networks

We study the problem of certifying the robustness of Bayesian neural net...
research
04/21/2020

Probabilistic Safety for Bayesian Neural Networks

We study probabilistic safety for Bayesian Neural Networks (BNNs) under ...
research
11/17/2022

SparseVLR: A Novel Framework for Verified Locally Robust Sparse Neural Networks Search

The compute-intensive nature of neural networks (NNs) limits their deplo...
research
09/17/2018

Robustness Guarantees for Bayesian Inference with Gaussian Processes

Bayesian inference and Gaussian processes are widely used in application...
research
04/01/2020

Tightened Convex Relaxations for Neural Network Robustness Certification

In this paper, we consider the problem of certifying the robustness of n...
research
10/29/2017

Certifiable Distributional Robustness with Principled Adversarial Training

Neural networks are vulnerable to adversarial examples and researchers h...

Please sign up or login with your details

Forgot password? Click here to reset