Periphery-Fovea Multi-Resolution Driving Model guided by Human Attention

03/24/2019
by   Ye Xia, et al.
24

Inspired by human vision, we propose a new periphery-fovea multi-resolution driving model that predicts vehicle speed from dash camera videos. The peripheral vision module of the model processes the full video frames in low resolution. Its foveal vision module selects sub-regions and uses high-resolution input from those regions to improve its driving performance. We train the fovea selection module with supervision from driver gaze. We show that adding high-resolution input from predicted human driver gaze locations significantly improves the driving accuracy of the model. Our periphery-fovea multi-resolution model outperforms a uni-resolution periphery-only model that has the same amount of floating-point operations. More importantly, we demonstrate that our driving model achieves a significantly higher performance gain in pedestrian-involved critical situations than in other non-critical situations.

READ FULL TEXT

page 1

page 3

page 4

page 6

research
11/17/2017

Training a network to attend like human drivers saves it from common but misleading loss functions

We proposed a novel FCN-ConvLSTM model to predict multi-focal human driv...
research
02/02/2021

Gaze-based dual resolution deep imitation learning for high-precision dexterous robot manipulation

A high-precision manipulation task, such as needle threading, is challen...
research
12/23/2020

Estimation of Driver's Gaze Region from Head Position and Orientation using Probabilistic Confidence Regions

A smart vehicle should be able to understand human behavior and predict ...
research
09/11/2023

HiLM-D: Towards High-Resolution Understanding in Multimodal Large Language Models for Autonomous Driving

Autonomous driving systems generally employ separate models for differen...
research
09/07/2019

Non-discriminative data or weak model? On the relative importance of data and model resolution

We explore the question of how the resolution of the input image ("input...
research
11/24/2019

"Looking at the right stuff" – Guided semantic-gaze for autonomous driving

In recent years, predicting driver's focus of attention has been a very ...
research
10/31/2019

A Self Validation Network for Object-Level Human Attention Estimation

Due to the foveated nature of the human vision system, people can focus ...

Please sign up or login with your details

Forgot password? Click here to reset