Improving Out-of-Distribution Robustness of Classifiers via Generative Interpolation

07/23/2023
by   Haoyue Bai, et al.
0

Deep neural networks achieve superior performance for learning from independent and identically distributed (i.i.d.) data. However, their performance deteriorates significantly when handling out-of-distribution (OoD) data, where the training and test are drawn from different distributions. In this paper, we explore utilizing the generative models as a data augmentation source for improving out-of-distribution robustness of neural classifiers. Specifically, we develop a simple yet effective method called Generative Interpolation to fuse generative models trained from multiple domains for synthesizing diverse OoD samples. Training a generative model directly on the source domains tends to suffer from mode collapse and sometimes amplifies the data bias. Instead, we first train a StyleGAN model on one source domain and then fine-tune it on the other domains, resulting in many correlated generators where their model parameters have the same initialization thus are aligned. We then linearly interpolate the model parameters of the generators to spawn new sets of generators. Such interpolated generators are used as an extra data augmentation source to train the classifiers. The interpolation coefficients can flexibly control the augmentation direction and strength. In addition, a style-mixing mechanism is applied to further improve the diversity of the generated OoD samples. Our experiments show that the proposed method explicitly increases the diversity of training domains and achieves consistent improvements over baselines across datasets and multiple different distribution shifts.

READ FULL TEXT

page 4

page 5

page 11

page 12

page 13

research
07/26/2023

Regularizing Neural Networks with Meta-Learning Generative Models

This paper investigates methods for improving generative data augmentati...
research
01/07/2022

GenLabel: Mixup Relabeling using Generative Models

Mixup is a data augmentation method that generates new data points by mi...
research
04/13/2020

Data augmentation using generative networks to identify dementia

Data limitation is one of the most common issues in training machine lea...
research
04/25/2022

VITA: A Multi-Source Vicinal Transfer Augmentation Method for Out-of-Distribution Generalization

Invariance to diverse types of image corruption, such as noise, blurring...
research
04/04/2023

PODIA-3D: Domain Adaptation of 3D Generative Model Across Large Domain Gap Using Pose-Preserved Text-to-Image Diffusion

Recently, significant advancements have been made in 3D generative model...
research
10/26/2021

AugMax: Adversarial Composition of Random Augmentations for Robust Training

Data augmentation is a simple yet effective way to improve the robustnes...
research
06/08/2023

Enhancing Robustness of AI Offensive Code Generators via Data Augmentation

In this work, we present a method to add perturbations to the code descr...

Please sign up or login with your details

Forgot password? Click here to reset