Semantic Equivariant Mixup

08/12/2023
by   Zongbo Han, et al.
0

Mixup is a well-established data augmentation technique, which can extend the training distribution and regularize the neural networks by creating ”mixed” samples based on the label-equivariance assumption, i.e., a proportional mixup of the input data results in the corresponding labels being mixed in the same proportion. However, previous mixup variants may fail to exploit the label-independent information in mixed samples during training, which usually contains richer semantic information. To further release the power of mixup, we first improve the previous label-equivariance assumption by the semantic-equivariance assumption, which states that the proportional mixup of the input data should lead to the corresponding representation being mixed in the same proportion. Then a generic mixup regularization at the representation level is proposed, which can further regularize the model with the semantic information in mixed samples. At a high level, the proposed semantic equivariant mixup (sem) encourages the structure of the input data to be preserved in the representation space, i.e., the change of input will result in the obtained representation information changing in the same way. Different from previous mixup variants, which tend to over-focus on the label-related information, the proposed method aims to preserve richer semantic information in the input with semantic-equivariance assumption, thereby improving the robustness of the model against distribution shifts. We conduct extensive empirical studies and qualitative analyzes to demonstrate the effectiveness of our proposed method. The code of the manuscript is in the supplement.

READ FULL TEXT
research
05/31/2022

Semantic Autoencoder and Its Potential Usage for Adversarial Attack

Autoencoder can give rise to an appropriate latent representation of the...
research
10/10/2019

First Order Ambisonics Domain Spatial Augmentation for DNN-based Direction of Arrival Estimation

In this paper, we propose a novel data augmentation method for training ...
research
11/04/2021

Stable and Compact Face Recognition via Unlabeled Data Driven Sparse Representation-Based Classification

Sparse representation-based classification (SRC) has attracted much atte...
research
10/04/2022

Nuisances via Negativa: Adjusting for Spurious Correlations via Data Augmentation

There exist features that are related to the label in the same way acros...
research
06/13/2018

Adversarial Learning with Local Coordinate Coding

Generative adversarial networks (GANs) aim to generate realistic data fr...
research
11/16/2019

Signed Input Regularization

Over-parameterized deep models usually over-fit to a given training dist...

Please sign up or login with your details

Forgot password? Click here to reset