A Unified Analysis of Mixed Sample Data Augmentation: A Loss Function Perspective

08/21/2022
by   Chanwoo Park, et al.
18

We propose the first unified theoretical analysis of mixed sample data augmentation (MSDA), such as Mixup and CutMix. Our theoretical results show that regardless of the choice of the mixing strategy, MSDA behaves as a pixel-level regularization of the underlying training loss and a regularization of the first layer parameters. Similarly, our theoretical results support that the MSDA training strategy can improve adversarial robustness and generalization compared to the vanilla training strategy. Using the theoretical results, we provide a high-level understanding of how different design choices of MSDA work differently. For example, we show that the most popular MSDA methods, Mixup and CutMix, behave differently, e.g., CutMix regularizes the input gradients by pixel distances, while Mixup regularizes the input gradients regardless of pixel distances. Our theoretical results also show that the optimal MSDA strategy depends on tasks, datasets, or model parameters. From these observations, we propose generalized MSDAs, a Hybrid version of Mixup and CutMix (HMix) and Gaussian Mixup (GMix), simple extensions of Mixup and CutMix. Our implementation can leverage the advantages of Mixup and CutMix, while our implementation is very efficient, and the computation cost is almost neglectable as Mixup and CutMix. Our empirical study shows that our HMix and GMix outperform the previous state-of-the-art MSDA methods in CIFAR-100 and ImageNet classification tasks. Source code is available at https://github.com/naver-ai/hmix-gmix

READ FULL TEXT

page 5

page 7

research
10/14/2022

TokenMixup: Efficient Attention-guided Token-level Data Augmentation for Transformers

Mixup is a commonly adopted data augmentation technique for image classi...
research
06/26/2022

Multiple Instance Learning with Mixed Supervision in Gleason Grading

With the development of computational pathology, deep learning methods f...
research
05/04/2023

LatentAugment: Dynamically Optimized Latent Probabilities of Data Augmentation

Although data augmentation is a powerful technique for improving the per...
research
02/27/2020

Understanding and Enhancing Mixed Sample Data Augmentation

Mixed Sample Data Augmentation (MSDA) has received increasing attention ...
research
01/21/2020

batchboost: regularization for stabilizing training with resistance to underfitting overfitting

Overfitting underfitting and stable training are an important challe...
research
05/27/2023

Toward Understanding Generative Data Augmentation

Generative data augmentation, which scales datasets by obtaining fake la...
research
01/15/2021

A Bayesian perspective on sampling of alternatives

In this paper, we apply a Bayesian perspective to sampling of alternativ...

Please sign up or login with your details

Forgot password? Click here to reset