Calibrated prediction in and out-of-domain for state-of-the-art saliency modeling

05/26/2021 ∙ by Akis Linardos, et al. ∙ 11

Since 2014 transfer learning has become the key driver for the improvement of spatial saliency prediction; however, with stagnant progress in the last 3-5 years. We conduct a large-scale transfer learning study which tests different ImageNet backbones, always using the same read out architecture and learning protocol adopted from DeepGaze II. By replacing the VGG19 backbone of DeepGaze II with ResNet50 features we improve the performance on saliency prediction from 78 backbones (such as EfficientNetB5) we observe no additional improvement on saliency prediction. By analyzing the backbones further, we find that generalization to other datasets differs substantially, with models being consistently overconfident in their fixation predictions. We show that by combining multiple backbones in a principled manner a good confidence calibration on unseen datasets can be achieved. This yields a significant leap in benchmark performance in and out-of-domain with a 15 percent point improvement over DeepGaze II to 93 on the MIT/Tuebingen Saliency Benchmark in all available metrics (AUC: 88.3 sAUC: 79.4



There are no comments yet.


page 1

page 3

page 4

page 7

page 14

Code Repositories


pytorch implementation of the different DeepGaze models

view repo
This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.