Constant Random Perturbations Provide Adversarial Robustness with Minimal Effect on Accuracy

03/15/2021
by   Bronya Roni Chernyak, et al.
0

This paper proposes an attack-independent (non-adversarial training) technique for improving adversarial robustness of neural network models, with minimal loss of standard accuracy. We suggest creating a neighborhood around each training example, such that the label is kept constant for all inputs within that neighborhood. Unlike previous work that follows a similar principle, we apply this idea by extending the training set with multiple perturbations for each training example, drawn from within the neighborhood. These perturbations are model independent, and remain constant throughout the entire training process. We analyzed our method empirically on MNIST, SVHN, and CIFAR-10, under different attacks and conditions. Results suggest that the proposed approach improves standard accuracy over other defenses while having increased robustness compared to vanilla adversarial training.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/14/2019

Confidence-Calibrated Adversarial Training: Towards Robust Models Generalizing Beyond the Attack Used During Training

Adversarial training is the standard to train models robust against adve...
research
02/25/2020

Understanding and Mitigating the Tradeoff Between Robustness and Accuracy

Adversarial training augments the training set with perturbations to imp...
research
09/10/2019

Localized Adversarial Training for Increased Accuracy and Robustness in Image Classification

Today's state-of-the-art image classifiers fail to correctly classify ca...
research
09/01/2021

Towards Improving Adversarial Training of NLP Models

Adversarial training, a method for learning robust deep neural networks,...
research
04/10/2018

Adversarial Training Versus Weight Decay

Performance-critical machine learning models should be robust to input p...
research
06/13/2022

Towards Alternative Techniques for Improving Adversarial Robustness: Analysis of Adversarial Training at a Spectrum of Perturbations

Adversarial training (AT) and its variants have spearheaded progress in ...
research
03/30/2020

Towards Deep Learning Models Resistant to Large Perturbations

Adversarial robustness has proven to be a required property of machine l...

Please sign up or login with your details

Forgot password? Click here to reset