How Drones Look: Crowdsourced Knowledge Transfer for Aerial Video Saliency Prediction

11/14/2018
by   Kui Fu, et al.
10

In ground-level platforms, many saliency models have been developed to perceive the visual world as the human does. However, they may not fit a drone that can look from many abnormal viewpoints. To address this problem, this paper proposes a Crowdsourced Multi-path Network (CMNet) that transfer the ground-level knowledge for spatiotemporal saliency prediction in aerial videos. To train CMNet, we first collect and fuse the eye-tracking data of 24 subjects on 1,000 aerial videos to annotate the ground-truth salient regions. Inspired by the crowdsourced annotation in eye-tracking experiments, we design a multi-path architecture for CMNet, in which each path is initialized under the supervision of a classic ground-level saliency model. After that, the most representative paths are selected in a data-driven manner, which are then fused and simultaneously fine-tuned on aerial videos. In this manner, the prior knowledge in various classic ground-level saliency models can be transferred into CMNet so as to improve its capability in processing aerial videos. Finally, the spatial predictions given by CMNet are adaptively refined by incorporating the temporal saliency predictions via a spatiotemporal saliency optimization algorithm. Experimental results show that the proposed approach outperforms ten state-of-the-art models in predicting visual saliency in aerial videos.

READ FULL TEXT

page 1

page 2

page 4

page 5

page 6

page 7

page 9

page 12

research
04/10/2019

Spatiotemporal Knowledge Distillation for Efficient Estimation of Aerial Video Saliency

The performance of video saliency estimation techniques has achieved sig...
research
07/27/2020

Saliency Prediction with External Knowledge

The last decades have seen great progress in saliency prediction, with t...
research
11/01/2016

A Benchmark Dataset and Saliency-guided Stacked Autoencoders for Video-based Salient Object Detection

Image-based salient object detection (SOD) has been extensively studied ...
research
01/06/2019

Unsupervised uncertainty estimation using spatiotemporal cues in video saliency detection

In this paper, we address the problem of quantifying reliability of comp...
research
01/09/2020

STAViS: Spatio-Temporal AudioVisual Saliency Network

We introduce STAViS, a spatio-temporal audiovisual saliency network that...
research
02/27/2020

A Neuromorphic Proto-Object Based Dynamic Visual Saliency Model with an FPGA Implementation

The ability to attend to salient regions of a visual scene is an innate ...
research
03/11/2016

Learning Gaze Transitions from Depth to Improve Video Saliency Estimation

In this paper we introduce a novel Depth-Aware Video Saliency approach t...

Please sign up or login with your details

Forgot password? Click here to reset