Enhanced Performance of Pre-Trained Networks by Matched Augmentation Distributions

01/19/2022
by   Touqeer Ahmad, et al.
0

There exists a distribution discrepancy between training and testing, in the way images are fed to modern CNNs. Recent work tried to bridge this gap either by fine-tuning or re-training the network at different resolutions. However re-training a network is rarely cheap and not always viable. To this end, we propose a simple solution to address the train-test distributional shift and enhance the performance of pre-trained models – which commonly ship as a package with deep learning platforms , PyTorch. Specifically, we demonstrate that running inference on the center crop of an image is not always the best as important discriminatory information may be cropped-off. Instead we propose to combine results for multiple random crops for a test image. This not only matches the train time augmentation but also provides the full coverage of the input image. We explore combining representation of random crops through averaging at different levels , deep feature level, logit level, and softmax level. We demonstrate that, for various families of modern deep networks, such averaging results in better validation accuracy compared to using a single central crop per image. The softmax averaging results in the best performance for various pre-trained networks without requiring any re-training or fine-tuning whatsoever. On modern GPUs with batch processing, the paper's approach to inference of pre-trained networks, is essentially free as all images in a batch can all be processed at once.

READ FULL TEXT

page 1

page 3

research
03/10/2022

Model soups: averaging weights of multiple fine-tuned models improves accuracy without increasing inference time

The conventional recipe for maximizing model accuracy is to (1) train mu...
research
03/09/2022

Inadequately Pre-trained Models are Better Feature Extractors

Pre-training has been a popular learning paradigm in deep learning era, ...
research
10/14/2020

Deep Ensembles for Low-Data Transfer Learning

In the low-data regime, it is difficult to train good supervised models ...
research
05/31/2021

HiddenCut: Simple Data Augmentation for Natural Language Understanding with Better Generalization

Fine-tuning large pre-trained models with task-specific data has achieve...
research
06/14/2019

Fixing the train-test resolution discrepancy

Data-augmentation is key to the training of neural networks for image cl...
research
02/02/2021

Test-Time Adaptation for Out-of-distributed Image Inpainting

Deep learning-based image inpainting algorithms have shown great perform...
research
04/07/2016

Families in the Wild (FIW): Large-Scale Kinship Image Database and Benchmarks

We present the largest kinship recognition dataset to date, Families in ...

Please sign up or login with your details

Forgot password? Click here to reset