MaxUp: A Simple Way to Improve Generalization of Neural Network Training

02/20/2020
by   Chengyue Gong, et al.
0

We propose MaxUp, an embarrassingly simple, highly effective technique for improving the generalization performance of machine learning models, especially deep neural networks. The idea is to generate a set of augmented data with some random perturbations or transforms and minimize the maximum, or worst case loss over the augmented data. By doing so, we implicitly introduce a smoothness or robustness regularization against the random perturbations, and hence improve the generation performance. For example, in the case of Gaussian perturbation, MaxUp is asymptotically equivalent to using the gradient norm of the loss as a penalty to encourage smoothness. We test MaxUp on a range of tasks, including image classification, language modeling, and adversarial certification, on which MaxUp consistently outperforms the existing best baseline methods, without introducing substantial computational overhead. In particular, we improve ImageNet classification from the state-of-the-art top-1 accuracy 85.5% without extra data to 85.8%. Code will be released soon.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/02/2021

Defending Against Image Corruptions Through Adversarial Augmentations

Modern neural networks excel at image classification, yet they remain vu...
research
02/20/2019

Graph Adversarial Training: Dynamically Regularizing Based on Graph Structure

Recent efforts show that neural networks are vulnerable to small but int...
research
12/14/2020

A case for new neural network smoothness constraints

How sensitive should machine learning models be to input changes? We tac...
research
05/17/2022

Monotonicity Regularization: Improved Penalties and Novel Applications to Disentangled Representation Learning and Robust Classification

We study settings where gradient penalties are used alongside risk minim...
research
04/09/2022

The Two Dimensions of Worst-case Training and the Integrated Effect for Out-of-domain Generalization

Training with an emphasis on "hard-to-learn" components of the data has ...
research
12/09/2022

Adversarial Weight Perturbation Improves Generalization in Graph Neural Network

A lot of theoretical and empirical evidence shows that the flatter local...
research
09/15/2021

ARCH: Efficient Adversarial Regularized Training with Caching

Adversarial regularization can improve model generalization in many natu...

Please sign up or login with your details

Forgot password? Click here to reset