Improving Adversarial Robustness of Ensembles with Diversity Training

01/28/2019
by   Sanjay Kariyappa, et al.
0

Deep Neural Networks are vulnerable to adversarial attacks even in settings where the attacker has no direct access to the model being attacked. Such attacks usually rely on the principle of transferability, whereby an attack crafted on a surrogate model tends to transfer to the target model. We show that an ensemble of models with misaligned loss gradients can provide an effective defense against transfer-based attacks. Our key insight is that an adversarial example is less likely to fool multiple models in the ensemble if their loss functions do not increase in a correlated fashion. To this end, we propose Diversity Training, a novel method to train an ensemble of models with uncorrelated loss functions. We show that our method significantly improves the adversarial robustness of ensembles and can also be combined with existing methods to create a stronger defense.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/09/2021

PARL: Enhancing Diversity of Ensemble Networks to Resist Adversarial Attacks via Pairwise Adversarially Robust Loss Function

The security of Deep Learning classifiers is a critical field of study b...
research
03/04/2023

Improved Robustness Against Adaptive Attacks With Ensembles and Error-Correcting Output Codes

Neural network ensembles have been studied extensively in the context of...
research
03/22/2023

Distribution-restrained Softmax Loss for the Model Robustness

Recently, the robustness of deep learning models has received widespread...
research
06/05/2021

Ensemble Defense with Data Diversity: Weak Correlation Implies Strong Robustness

In this paper, we propose a framework of filter-based ensemble of deep n...
research
01/25/2019

Improving Adversarial Robustness via Promoting Ensemble Diversity

Though deep neural networks have achieved significant progress on variou...
research
07/06/2021

GradDiv: Adversarial Robustness of Randomized Neural Networks via Gradient Diversity Regularization

Deep learning is vulnerable to adversarial examples. Many defenses based...
research
10/06/2022

Dynamic Stochastic Ensemble with Adversarial Robust Lottery Ticket Subnetworks

Adversarial attacks are considered the intrinsic vulnerability of CNNs. ...

Please sign up or login with your details

Forgot password? Click here to reset