CC-Cert: A Probabilistic Approach to Certify General Robustness of Neural Networks

09/22/2021
by   Mikhail Pautov, et al.
0

In safety-critical machine learning applications, it is crucial to defend models against adversarial attacks – small modifications of the input that change the predictions. Besides rigorously studied ℓ_p-bounded additive perturbations, recently proposed semantic perturbations (e.g. rotation, translation) raise a serious concern on deploying ML systems in real-world. Therefore, it is important to provide provable guarantees for deep learning models against semantically meaningful input transformations. In this paper, we propose a new universal probabilistic certification approach based on Chernoff-Cramer bounds that can be used in general attack settings. We estimate the probability of a model to fail if the attack is sampled from a certain distribution. Our theoretical findings are supported by experimental results on different datasets.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/05/2022

PRoA: A Probabilistic Robustness Assessment against Functional Perturbations

In safety-critical deep learning applications robustness measurement is ...
research
08/15/2022

Man-in-the-Middle Attack against Object Detection Systems

Is deep learning secure for robots? As embedded systems have access to m...
research
06/22/2022

Robust Universal Adversarial Perturbations

Universal Adversarial Perturbations (UAPs) are imperceptible, image-agno...
research
09/19/2020

Efficient Certification of Spatial Robustness

Recent work has exposed the vulnerability of computer vision models to s...
research
03/28/2023

Provable Robustness for Streaming Models with a Sliding Window

The literature on provable robustness in machine learning has primarily ...
research
12/06/2019

Achieving Robustness in the Wild via Adversarial Mixing with Disentangled Representations

Recent research has made the surprising finding that state-of-the-art de...
research
04/21/2022

Robustness of Machine Learning Models Beyond Adversarial Attacks

Correctly quantifying the robustness of machine learning models is a cen...

Please sign up or login with your details

Forgot password? Click here to reset