Over-parameterized Adversarial Training: An Analysis Overcoming the Curse of Dimensionality

02/16/2020
by   Yi Zhang, et al.
11

Adversarial training is a popular method to give neural nets robustness against adversarial perturbations. In practice adversarial training leads to low robust training loss. However, a rigorous explanation for why this happens under natural conditions is still missing. Recently a convergence theory for standard (non-adversarial) supervised training was developed by various groups for very overparametrized nets. It is unclear how to extend these results to adversarial training because of the min-max objective. Recently, a first step towards this direction was made by Gao et al. using tools from online learning, but they require the width of the net to be exponential in input dimension d, and with an unnatural activation function. Our work proves convergence to low robust training loss for polynomial width instead of exponential, under natural assumptions and with the ReLU activation. Key element of our proof is showing that ReLU networks near initialization can approximate the step function, which may be of independent interest.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/10/2022

A Sublinear Adversarial Training Algorithm

Adversarial training is a widely used strategy for making neural network...
research
05/20/2020

Feature Purification: How Adversarial Training Performs Robust Deep Learning

Despite the great empirical success of adversarial training to defend de...
research
06/25/2020

Smooth Adversarial Training

It is commonly believed that networks cannot be both accurate and robust...
research
06/13/2022

Distributed Adversarial Training to Robustify Deep Neural Networks at Scale

Current deep neural networks (DNNs) are vulnerable to adversarial attack...
research
06/19/2019

Convergence of Adversarial Training in Overparametrized Networks

Neural networks are vulnerable to adversarial examples, i.e. inputs that...
research
02/15/2021

Low Curvature Activations Reduce Overfitting in Adversarial Training

Adversarial training is one of the most effective defenses against adver...
research
10/11/2021

Parameterizing Activation Functions for Adversarial Robustness

Deep neural networks are known to be vulnerable to adversarially perturb...

Please sign up or login with your details

Forgot password? Click here to reset