Improving Adversarial Robustness by Enforcing Local and Global Compactness

07/10/2020
by   Anh Bui, et al.
7

The fact that deep neural networks are susceptible to crafted perturbations severely impacts the use of deep learning in certain domains of application. Among many developed defense models against such attacks, adversarial training emerges as the most successful method that consistently resists a wide range of attacks. In this work, based on an observation from a previous study that the representations of a clean data example and its adversarial examples become more divergent in higher layers of a deep neural net, we propose the Adversary Divergence Reduction Network which enforces local/global compactness and the clustering assumption over an intermediate layer of a deep neural network. We conduct comprehensive experiments to understand the isolating behavior of each component (i.e., local/global compactness and the clustering assumption) and compare our proposed model with state-of-the-art adversarial training methods. The experimental results demonstrate that augmenting adversarial training with our proposed components can further improve the robustness of the network, leading to higher unperturbed and adversarial predictive performances.

READ FULL TEXT
research
01/12/2018

A3T: Adversarially Augmented Adversarial Training

Recent research showed that deep neural networks are highly sensitive to...
research
05/10/2020

Class-Aware Domain Adaptation for Improving Adversarial Robustness

Recent works have demonstrated convolutional neural networks are vulnera...
research
06/13/2020

ClustTR: Clustering Training for Robustness

This paper studies how encouraging semantically-aligned features during ...
research
12/08/2020

Overcomplete Representations Against Adversarial Videos

Adversarial robustness of deep neural networks is an extensively studied...
research
07/21/2021

Fast and Scalable Adversarial Training of Kernel SVM via Doubly Stochastic Gradients

Adversarial attacks by generating examples which are almost indistinguis...
research
09/20/2019

Defending Against Physically Realizable Attacks on Image Classification

We study the problem of defending deep neural network approaches for ima...
research
06/10/2019

Intriguing properties of adversarial training

Adversarial training is one of the main defenses against adversarial att...

Please sign up or login with your details

Forgot password? Click here to reset