Denoising Autoencoder-based Defensive Distillation as an Adversarial Robustness Algorithm

03/28/2023
by   Bakary Badjie, et al.
0

Adversarial attacks significantly threaten the robustness of deep neural networks (DNNs). Despite the multiple defensive methods employed, they are nevertheless vulnerable to poison attacks, where attackers meddle with the initial training data. In order to defend DNNs against such adversarial attacks, this work proposes a novel method that combines the defensive distillation mechanism with a denoising autoencoder (DAE). This technique tries to lower the sensitivity of the distilled model to poison attacks by spotting and reconstructing poisonous adversarial inputs in the training data. We added carefully created adversarial samples to the initial training data to assess the proposed method's performance. Our experimental findings demonstrate that our method successfully identified and reconstructed the poisonous inputs while also considering enhancing the DNN's resilience. The proposed approach provides a potent and robust defense mechanism for DNNs in various applications where data poisoning attacks are a concern. Thus, the defensive distillation technique's limitation posed by poisonous adversarial attacks is overcome.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/05/2020

Bluff: Interactively Deciphering Adversarial Attacks on Deep Neural Networks

Deep neural networks (DNNs) are now commonly used in many domains. Howev...
research
11/04/2018

QuSecNets: Quantization-based Defense Mechanism for Securing Deep Neural Network against Adversarial Attacks

Deep Neural Networks (DNNs) have recently been shown vulnerable to adver...
research
01/08/2021

More Tolerant Reconstructed Networks by Self-Healing against Attacks in Saving Resource

Complex network infrastructure systems for power-supply, communication, ...
research
08/21/2022

MockingBERT: A Method for Retroactively Adding Resilience to NLP Models

Protecting NLP models against misspellings whether accidental or adversa...
research
11/14/2015

Distillation as a Defense to Adversarial Perturbations against Deep Neural Networks

Deep learning algorithms have been shown to perform extremely well on ma...
research
06/08/2022

Can Backdoor Attacks Survive Time-Varying Models?

Backdoors are powerful attacks against deep neural networks (DNNs). By p...

Please sign up or login with your details

Forgot password? Click here to reset