Phase-shifted Adversarial Training

01/12/2023
by   Yeachan Kim, et al.
0

Adversarial training has been considered an imperative component for safely deploying neural network-based applications to the real world. To achieve stronger robustness, existing methods primarily focus on how to generate strong attacks by increasing the number of update steps, regularizing the models with the smoothed loss function, and injecting the randomness into the attack. Instead, we analyze the behavior of adversarial training through the lens of response frequency. We empirically discover that adversarial training causes neural networks to have low convergence to high-frequency information, resulting in highly oscillated predictions near each data. To learn high-frequency contents efficiently and effectively, we first prove that a universal phenomenon of frequency principle, i.e., lower frequencies are learned first, still holds in adversarial training. Based on that, we propose phase-shifted adversarial training (PhaseAT) in which the model learns high-frequency components by shifting these frequencies to the low-frequency range where the fast convergence occurs. For evaluations, we conduct the experiments on CIFAR-10 and ImageNet with the adaptive attack carefully designed for reliable evaluation. Comprehensive results show that PhaseAT significantly improves the convergence for high-frequency information. This results in improved adversarial robustness by enabling the model to have smoothed predictions near each data.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/19/2019

Adversarial Defense by Suppressing High-frequency Components

Recent works show that deep neural networks trained on image classificat...
research
10/13/2020

Toward Few-step Adversarial Training from a Frequency Perspective

We investigate adversarial-sample generation methods from a frequency do...
research
02/17/2023

High-frequency Matters: An Overwriting Attack and defense for Image-processing Neural Network Watermarking

In recent years, there has been significant advancement in the field of ...
research
05/21/2023

Generative Model Watermarking Suppressing High-Frequency Artifacts

Protecting deep neural networks (DNNs) against intellectual property (IP...
research
07/19/2023

Towards Building More Robust Models with Frequency Bias

The vulnerability of deep neural networks to adversarial samples has bee...
research
02/28/2019

On the Effectiveness of Low Frequency Perturbations

Carefully crafted, often imperceptible, adversarial perturbations have b...
research
05/06/2020

Towards Frequency-Based Explanation for Robust CNN

Current explanation techniques towards a transparent Convolutional Neura...

Please sign up or login with your details

Forgot password? Click here to reset