Learning to Attend Relevant Regions in Videos from Eye Fixations

11/21/2018
by   Thanh T. Nguyen, et al.
0

Attentively important objects in videos account for a majority part of semantics in a current frame. Information about human attention might be useful not only for entertainment (such as auto generating commentary and tourist guide) but also for robotic control which holds a larascope supported for laparoscopic surgery. In this work, we address the problem of attending relevant objects in videos conditioned on eye fixations using RNN-based visual attention model. To the best of our knowledge, this is the first work to approach the problem from RNNs.

READ FULL TEXT

page 5

page 6

research
01/30/2019

Understanding spatial correlation in eye-fixation maps for visual attention in videos

In this paper, we present an analysis of recorded eye-fixation data from...
research
02/23/2018

A Weighted Sparse Sampling and Smoothing Frame Transition Approach for Semantic Fast-Forward First-Person Videos

Thanks to the advances in the technology of low-cost digital cameras and...
research
05/22/2018

Teacher's Perception in the Classroom

The ability for a teacher to engage all students in active learning proc...
research
04/03/2020

Privacy-Preserving Eye Videos using Rubber Sheet Model

Video-based eye trackers estimate gaze based on eye images/videos. As se...
research
03/05/2020

Detecting Attended Visual Targets in Video

We address the problem of detecting attention targets in video. Specific...
research
09/21/2020

A Sparse Sampling-based framework for Semantic Fast-Forward of First-Person Videos

Technological advances in sensors have paved the way for digital cameras...
research
05/24/2019

Overt visual attention on rendered 3D objects

This work covers multiple aspects of overt visual attention on 3D render...

Please sign up or login with your details

Forgot password? Click here to reset