Do Better ImageNet Models Transfer Better?

05/23/2018
by   Simon Kornblith, et al.
0

Transfer learning has become a cornerstone of computer vision with the advent of ImageNet features, yet little work has been done to evaluate the performance of ImageNet architectures across different datasets. An implicit hypothesis in modern computer vision research is that models that perform better on ImageNet necessarily perform better on other vision tasks. However, this hypothesis has never been systematically tested. Here, we compare the performance of 13 classification models on 12 image classification tasks in three settings: as fixed feature extractors, fine-tuned, and trained from random initialization. We find that, when networks are used as fixed feature extractors, ImageNet accuracy is only weakly predictive of accuracy on other tasks (r^2=0.24). In this setting, ResNets consistently outperform networks that achieve higher accuracy on ImageNet. When networks are fine-tuned, we observe a substantially stronger correlation (r^2 = 0.86). We achieve state-of-the-art performance on eight image classification tasks simply by fine-tuning state-of-the-art ImageNet architectures, outperforming previous results based on specialized methods for transfer learning. Finally, we observe that, on three small fine-grained image classification datasets, networks trained from random initialization perform similarly to ImageNet-pretrained networks. Together, our results show that ImageNet architectures generalize well across datasets, with small improvements in ImageNet accuracy producing improvements across other tasks, but ImageNet features are less general than previously suggested.

READ FULL TEXT

page 2

page 6

research
07/25/2018

Do Better ImageNet Models Transfer Better... for Image Recommendation ?

Visual embeddings from Convolutional Neural Networks (CNN) trained on th...
research
03/16/2021

Is it Enough to Optimize CNN Architectures on ImageNet?

An implicit but pervasive hypothesis of modern computer vision research ...
research
07/22/2020

Rethinking CNN Models for Audio Classification

In this paper, we show that ImageNet-Pretrained standard deep CNN models...
research
02/13/2023

A Comprehensive Study of Modern Architectures and Regularization Approaches on CheXpert5000

Computer aided diagnosis (CAD) has gained an increased amount of attenti...
research
06/13/2021

HistoTransfer: Understanding Transfer Learning for Histopathology

Advancement in digital pathology and artificial intelligence has enabled...
research
05/26/2022

TransBoost: Improving the Best ImageNet Performance using Deep Transduction

This paper deals with deep transductive learning, and proposes TransBoos...

Please sign up or login with your details

Forgot password? Click here to reset