Explaining Autonomous Driving Actions with Visual Question Answering

07/19/2023
by   Shahin Atakishiyev, et al.
0

The end-to-end learning ability of self-driving vehicles has achieved significant milestones over the last decade owing to rapid advances in deep learning and computer vision algorithms. However, as autonomous driving technology is a safety-critical application of artificial intelligence (AI), road accidents and established regulatory principles necessitate the need for the explainability of intelligent action choices for self-driving vehicles. To facilitate interpretability of decision-making in autonomous driving, we present a Visual Question Answering (VQA) framework, which explains driving actions with question-answering-based causal reasoning. To do so, we first collect driving videos in a simulation environment using reinforcement learning (RL) and extract consecutive frames from this log data uniformly for five selected action categories. Further, we manually annotate the extracted frames using question-answer pairs as justifications for the actions chosen in each scenario. Finally, we evaluate the correctness of the VQA-predicted answers for actions on unseen driving scenes. The empirical results suggest that the VQA mechanism can provide support to interpret real-time decisions of autonomous vehicles and help enhance overall driving safety.

READ FULL TEXT

page 1

page 3

page 6

research
05/31/2019

Out of Sight But Not Out of Mind: An Answer Set Programming Based Online Abduction Framework for Visual Sensemaking in Autonomous Driving

We demonstrate the need and potential of systematically integrated visio...
research
07/18/2023

Towards a performance analysis on pre-trained Visual Question Answering models for autonomous driving

This short paper presents a preliminary analysis of three popular Visual...
research
12/21/2021

Explainable Artificial Intelligence for Autonomous Driving: A Comprehensive Overview and Field Guide for Future Research Directions

Autonomous driving has achieved a significant milestone in research and ...
research
05/24/2023

NuScenes-QA: A Multi-modal Visual Question Answering Benchmark for Autonomous Driving Scenario

We introduce a novel visual question answering (VQA) task in the context...
research
12/28/2020

Commonsense Visual Sensemaking for Autonomous Driving: On Generalised Neurosymbolic Online Abduction Integrating Vision and Semantics

We demonstrate the need and potential of systematically integrated visio...
research
11/27/2020

The NEOLIX Open Dataset for AutonomousDriving

With the gradual maturity of 5G technology,autonomous driving technology...
research
11/08/2022

SOTIF Entropy: Online SOTIF Risk Quantification and Mitigation for Autonomous Driving

Autonomous driving confronts great challenges in complex traffic scenari...

Please sign up or login with your details

Forgot password? Click here to reset