Using mixup as regularization and tuning hyper-parameters for ResNets

While novel computer vision architectures are gaining traction, the impact of model architectures is often related to changes or exploring in training methods. Identity mapping-based architectures ResNets and DenseNets have promised path-breaking results in the image classification task and are go-to methods for even now if the data given is fairly limited. Considering the ease of training with limited resources this work revisits the ResNets and improves the ResNet50 <cit.> by using mixup data-augmentation as regularization and tuning the hyper-parameters.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/16/2021

Towards Efficient and Data Agnostic Image Classification Training Pipeline for Embedded Systems

Nowadays deep learning-based methods have achieved a remarkable progress...
research
12/16/2021

How to augment your ViTs? Consistency loss and StyleAug, a random style transfer augmentation

The Vision Transformer (ViT) architecture has recently achieved competit...
research
09/04/2023

No Data Augmentation? Alternative Regularizations for Effective Training on Small Datasets

Solving image classification tasks given small training datasets remains...
research
01/31/2022

Plug-In Inversion: Model-Agnostic Inversion for Vision with Data Augmentations

Existing techniques for model inversion typically rely on hard-to-tune r...
research
11/09/2022

Hyper-Parameter Auto-Tuning for Sparse Bayesian Learning

Choosing the values of hyper-parameters in sparse Bayesian learning (SBL...
research
01/12/2021

Mixup Without Hesitation

Mixup linearly interpolates pairs of examples to form new samples, which...
research
08/07/2021

Impact of Aliasing on Generalization in Deep Convolutional Networks

We investigate the impact of aliasing on generalization in Deep Convolut...

Please sign up or login with your details

Forgot password? Click here to reset