Provable Robustness of Adversarial Training for Learning Halfspaces with Noise

04/19/2021
by   Difan Zou, et al.
8

We analyze the properties of adversarial training for learning adversarially robust halfspaces in the presence of agnostic label noise. Denoting 𝖮𝖯𝖳_p,r as the best robust classification error achieved by a halfspace that is robust to perturbations of ℓ_p balls of radius r, we show that adversarial training on the standard binary cross-entropy loss yields adversarially robust halfspaces up to (robust) classification error Õ(√(𝖮𝖯𝖳_2,r)) for p=2, and Õ(d^1/4√(𝖮𝖯𝖳_∞, r) + d^1/2𝖮𝖯𝖳_∞,r) when p=∞. Our results hold for distributions satisfying anti-concentration properties enjoyed by log-concave isotropic distributions among others. We additionally show that if one instead uses a nonconvex sigmoidal loss, adversarial training yields halfspaces with an improved robust classification error of O(𝖮𝖯𝖳_2,r) for p=2, and O(d^1/4𝖮𝖯𝖳_∞, r) when p=∞. To the best of our knowledge, this is the first work to show that adversarial training provably yields robust classifiers in the presence of noise.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/28/2023

On the existence of solutions to adversarial training in multiclass classification

We study three models of the problem of adversarial training in multicla...
research
11/19/2021

Fooling Adversarial Training with Inducing Noise

Adversarial training is widely believed to be a reliable approach to imp...
research
06/13/2023

On Achieving Optimal Adversarial Test Error

We first elucidate various fundamental properties of optimal adversarial...
research
10/07/2021

Double Descent in Adversarial Training: An Implicit Label Noise Perspective

Here, we show that the robust overfitting shall be viewed as the early p...
research
05/20/2020

Feature Purification: How Adversarial Training Performs Robust Deep Learning

Despite the great empirical success of adversarial training to defend de...
research
01/08/2020

MACER: Attack-free and Scalable Robust Training via Maximizing Certified Radius

Adversarial training is one of the most popular ways to learn robust mod...
research
03/03/2023

Certified Robust Neural Networks: Generalization and Corruption Resistance

Adversarial training aims to reduce the problematic susceptibility of mo...

Please sign up or login with your details

Forgot password? Click here to reset