Culture-Based Explainable Human-Agent Deconfliction

11/22/2019
by   Alex Raymond, et al.
0

Law codes and regulations help organise societies for centuries, and as AI systems gain more autonomy, we question how human-agent systems can operate as peers under the same norms, especially when resources are contended. We posit that agents must be accountable and explainable by referring to which rules justify their decisions. The need for explanations is associated with user acceptance and trust. This paper's contribution is twofold: i) we propose an argumentation-based human-agent architecture to map human regulations into a culture for artificial agents with explainable behaviour. Our architecture leans on the notion of argumentative dialogues and generates explanations from the history of such dialogues; and ii) we validate our architecture with a user study in the context of human-agent path deconfliction. Our results show that explanations provide a significantly higher improvement in human performance when systems are more complex. Consequently, we argue that the criteria defining the need of explanations should also consider the complexity of a system. Qualitative findings show that when rules are more complex, explanations significantly reduce the perception of challenge for humans.

READ FULL TEXT
research
05/05/2020

Don't Explain without Verifying Veracity: An Evaluation of Explainable AI with Video Activity Recognition

Explainable machine learning and artificial intelligence models have bee...
research
04/22/2023

Trust and Reliance in Consensus-Based Explanations from an Anti-Misinformation Agent

The illusion of consensus occurs when people believe there is consensus ...
research
12/19/2019

Interestingness Elements for Explainable Reinforcement Learning: Understanding Agents' Capabilities and Limitations

We propose an explainable reinforcement learning (XRL) framework that an...
research
03/27/2023

Interactive Explanations by Conflict Resolution via Argumentative Exchanges

As the field of explainable AI (XAI) is maturing, calls for interactive ...
research
07/02/2020

The Impact of Explanations on AI Competency Prediction in VQA

Explainability is one of the key elements for building trust in AI syste...
research
05/17/2023

Explainable Multi-Agent Reinforcement Learning for Temporal Queries

As multi-agent reinforcement learning (MARL) systems are increasingly de...
research
03/09/2023

Explainable Goal Recognition: A Framework Based on Weight of Evidence

We introduce and evaluate an eXplainable Goal Recognition (XGR) model th...

Please sign up or login with your details

Forgot password? Click here to reset