SPADE: A Spectral Method for Black-Box Adversarial Robustness Evaluation

02/07/2021
by   Wuxinlin Cheng, et al.
0

A black-box spectral method is introduced for evaluating the adversarial robustness of a given machine learning (ML) model. Our approach, named SPADE, exploits bijective distance mapping between the input/output graphs constructed for approximating the manifolds corresponding to the input/output data. By leveraging the generalized Courant-Fischer theorem, we propose a SPADE score for evaluating the adversarial robustness of a given model, which is proved to be an upper bound of the best Lipschitz constant under the manifold setting. To reveal the most non-robust data samples highly vulnerable to adversarial attacks, we develop a spectral graph embedding procedure leveraging dominant generalized eigenvectors. This embedding step allows assigning each data sample a robustness score that can be further harnessed for more effective adversarial training. Our experiments show the proposed SPADE method leads to promising empirical results for neural network models adversarially trained with the MNIST and CIFAR-10 data sets.

READ FULL TEXT

page 4

page 18

research
08/06/2018

Gray-box Adversarial Training

Adversarial samples are perturbed inputs crafted to mislead the machine ...
research
05/23/2019

Thwarting finite difference adversarial attacks with output randomization

Adversarial examples pose a threat to deep neural network models in a va...
research
02/04/2022

Pixle: a fast and effective black-box attack based on rearranging pixels

Recent research has found that neural networks are vulnerable to several...
research
12/01/2020

Robustness Out of the Box: Compositional Representations Naturally Defend Against Black-Box Patch Attacks

Patch-based adversarial attacks introduce a perceptible but localized ch...
research
04/30/2021

ModelGuard: Runtime Validation of Lipschitz-continuous Models

This paper presents ModelGuard, a sampling-based approach to runtime mod...
research
07/05/2018

Explainable Learning: Implicit Generative Modelling during Training for Adversarial Robustness

We introduce Explainable Learning ,ExL, an approach for training neural ...
research
05/09/2020

Estimating g-Leakage via Machine Learning

This paper considers the problem of estimating the information leakage o...

Please sign up or login with your details

Forgot password? Click here to reset