Deep Repulsive Prototypes for Adversarial Robustness

05/26/2021
by   Alex Serban, et al.
0

While many defences against adversarial examples have been proposed, finding robust machine learning models is still an open problem. The most compelling defence to date is adversarial training and consists of complementing the training data set with adversarial examples. Yet adversarial training severely impacts training time and depends on finding representative adversarial samples. In this paper we propose to train models on output spaces with large class separation in order to gain robustness without adversarial training. We introduce a method to partition the output space into class prototypes with large separation and train models to preserve it. Experimental results shows that models trained with these prototypes – which we call deep repulsive prototypes – gain robustness competitive with adversarial training, while also preserving more accuracy on natural samples. Moreover, the models are more resilient to large perturbation sizes. For example, we obtained over 50 robustness for CIFAR-10, with 92 robustness for CIFAR-100, with 71 adversarial training. For both data sets, the models preserved robustness against large perturbations better than adversarially trained models.

READ FULL TEXT
research
05/15/2019

On Norm-Agnostic Robustness of Adversarial Training

Adversarial examples are carefully perturbed in-puts for fooling machine...
research
10/13/2020

To be Robust or to be Fair: Towards Fairness in Adversarial Training

Adversarial training algorithms have been proven to be reliable to impro...
research
07/01/2022

Efficient Adversarial Training With Data Pruning

Neural networks are susceptible to adversarial examples-small input pert...
research
08/12/2020

Learning to Learn from Mistakes: Robust Optimization for Adversarial Noise

Sensitivity to adversarial noise hinders deployment of machine learning ...
research
10/16/2021

Analyzing Dynamic Adversarial Training Data in the Limit

To create models that are robust across a wide range of test inputs, tra...
research
02/01/2021

Towards Speeding up Adversarial Training in Latent Spaces

Adversarial training is wildly considered as the most effective way to d...
research
11/10/2022

Impact of Adversarial Training on Robustness and Generalizability of Language Models

Adversarial training is widely acknowledged as the most effective defens...

Please sign up or login with your details

Forgot password? Click here to reset