Adversarial Training with Generated Data in High-Dimensional Regression: An Asymptotic Study

06/21/2023
by   Yue Xing, et al.
0

In recent years, studies such as <cit.> have demonstrated that incorporating additional real or generated data with pseudo-labels can enhance adversarial training through a two-stage training approach. In this paper, we perform a theoretical analysis of the asymptotic behavior of this method in high-dimensional linear regression. While a double-descent phenomenon can be observed in ridgeless training, with an appropriate ℒ_2 regularization, the two-stage adversarial training achieves a better performance. Finally, we derive a shortcut cross-validation formula specifically tailored for the two-stage training method.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/26/2020

Asymptotic Behavior of Adversarial Training in Binary Classification

It is widely known that several machine learning models are susceptible ...
research
08/15/2020

On the Generalization Properties of Adversarial Training

Modern machine learning and deep learning models are shown to be vulnera...
research
05/25/2022

Surprises in adversarially-trained linear regression

State-of-the-art machine learning models can be vulnerable to very small...
research
03/03/2022

Why adversarial training can hurt robust accuracy

Machine learning classifiers with high test accuracy often perform poorl...
research
11/21/2021

Local Linearity and Double Descent in Catastrophic Overfitting

Catastrophic overfitting is a phenomenon observed during Adversarial Tra...
research
04/13/2023

Understanding Overfitting in Adversarial Training via Kernel Regression

Adversarial training and data augmentation with noise are widely adopted...
research
01/11/2023

Multivariate Regression via Enhanced Response Envelope: Envelope Regularization and Double Descent

The envelope model provides substantial efficiency gains over the standa...

Please sign up or login with your details

Forgot password? Click here to reset