MixMo: Mixing Multiple Inputs for Multiple Outputs via Deep Subnetworks

03/10/2021
by   Alexandre Ramé, et al.
0

Recent strategies achieved ensembling "for free" by fitting concurrently diverse subnetworks inside a single base network. The main idea during training is that each subnetwork learns to classify only one of the multiple inputs simultaneously provided. However, the question of how to best mix these multiple inputs has not been studied so far. In this paper, we introduce MixMo, a new generalized framework for learning multi-input multi-output deep subnetworks. Our key motivation is to replace the suboptimal summing operation hidden in previous approaches by a more appropriate mixing mechanism. For that purpose, we draw inspiration from successful mixed sample data augmentations. We show that binary mixing in features - particularly with rectangular patches from CutMix - enhances results by making subnetworks stronger and more diverse. We improve state of the art for image classification on CIFAR-100 and Tiny ImageNet datasets. Our easy to implement models notably outperform data augmented deep ensembles, without the inference and memory overheads. As we operate in features and simply better leverage the expressiveness of large networks, we open a new line of research complementary to previous works.

READ FULL TEXT
research
11/28/2017

Between-class Learning for Image Classification

In this paper, we propose a novel learning method for image classificati...
research
08/05/2023

MiAMix: Enhancing Image Classification through a Multi-stage Augmented Mixied Sample Data Augmentation Method

Despite substantial progress in the field of deep learning, overfitting ...
research
06/14/2020

PatchUp: A Regularization Technique for Convolutional Neural Networks

Large capacity deep learning models are often prone to a high generaliza...
research
12/16/2021

Saliency Grafting: Innocuous Attribution-Guided Mixup with Calibrated Label Mixing

The Mixup scheme suggests mixing a pair of samples to create an augmente...
research
10/08/2021

Observations on K-image Expansion of Image-Mixing Augmentation for Classification

Image-mixing augmentations (e.g., Mixup or CutMix), which typically mix ...
research
06/17/2021

ShuffleBlock: Shuffle to Regularize Deep Convolutional Neural Networks

Deep neural networks have enormous representational power which leads th...
research
05/05/2021

MOS: Towards Scaling Out-of-distribution Detection for Large Semantic Space

Detecting out-of-distribution (OOD) inputs is a central challenge for sa...

Please sign up or login with your details

Forgot password? Click here to reset