Domain Generalization via Balancing Training Difficulty and Model Capability

09/02/2023
by   Xueying Jiang, et al.
0

Domain generalization (DG) aims to learn domain-generalizable models from one or multiple source domains that can perform well in unseen target domains. Despite its recent progress, most existing work suffers from the misalignment between the difficulty level of training samples and the capability of contemporarily trained models, leading to over-fitting or under-fitting in the trained generalization model. We design MoDify, a Momentum Difficulty framework that tackles the misalignment by balancing the seesaw between the model's capability and the samples' difficulties along the training process. MoDify consists of two novel designs that collaborate to fight against the misalignment while learning domain-generalizable models. The first is MoDify-based Data Augmentation which exploits an RGB Shuffle technique to generate difficulty-aware training samples on the fly. The second is MoDify-based Network Optimization which dynamically schedules the training samples for balanced and smooth learning with appropriate difficulty. Without bells and whistles, a simple implementation of MoDify achieves superior performance across multiple benchmarks. In addition, MoDify can complement existing methods as a plug-in, and it is generic and can work for different visual recognition tasks.

READ FULL TEXT
research
09/11/2020

Heterogeneous Domain Generalization via Domain Mixup

One of the main drawbacks of deep Convolutional Neural Networks (DCNN) i...
research
01/12/2023

Understanding Difficulty-based Sample Weighting with a Universal Difficulty Measure

Sample weighting is widely used in deep learning. A large number of weig...
research
08/28/2023

LatentDR: Improving Model Generalization Through Sample-Aware Latent Degradation and Restoration

Despite significant advances in deep learning, models often struggle to ...
research
09/29/2022

Effective Vision Transformer Training: A Data-Centric Perspective

Vision Transformers (ViTs) have shown promising performance compared wit...
research
11/17/2022

UPTON: Unattributable Authorship Text via Data Poisoning

In online medium such as opinion column in Bloomberg, The Guardian and W...
research
10/24/2019

Superposition as Data Augmentation using LSTM and HMM in Small Training Sets

Considering audio and image data as having quantum nature (data are repr...
research
06/26/2023

Pseudo-Trilateral Adversarial Training for Domain Adaptive Traversability Prediction

Traversability prediction is a fundamental perception capability for aut...

Please sign up or login with your details

Forgot password? Click here to reset