Improved Explainability of Capsule Networks: Relevance Path by Agreement

02/27/2018
by   Atefeh Shahroudnejad, et al.
0

Recent advancements in signal processing and machine learning domains have resulted in an extensive surge of interest in deep learning models due to their unprecedented performance and high accuracy for different and challenging problems of significant engineering importance. However, when such deep learning architectures are utilized for making critical decisions such as the ones that involve human lives (e.g., in medical applications), it is of paramount importance to understand, trust, and in one word "explain" the rational behind deep models' decisions. Currently, deep learning models are typically considered as black-box systems, which do not provide any clue on their internal processing actions. Although some recent efforts have been initiated to explain behavior and decisions of deep networks, explainable artificial intelligence (XAI) domain is still in its infancy. In this regard, we consider capsule networks (referred to as CapsNets), which are novel deep structures; recently proposed as an alternative counterpart to convolutional neural networks (CNNs), and posed to change the future of machine intelligence. In this paper, we investigate and analyze structures and behaviors of the CapsNets and illustrate potential explainability properties of such networks. Furthermore, we show possibility of transforming deep learning architectures in to transparent networks via incorporation of capsules in different layers instead of convolution layers of the CNNs.

READ FULL TEXT
research
02/02/2021

A Survey on Understanding, Visualizations, and Explanation of Deep Neural Networks

Recent advancements in machine learning and signal processing domains ha...
research
08/25/2022

Towards Benchmarking Explainable Artificial Intelligence Methods

The currently dominating artificial intelligence and machine learning te...
research
02/06/2023

Stop overkilling simple tasks with black-box models and use transparent models instead

In recent years, the employment of deep learning methods has led to seve...
research
04/07/2018

Visual Analytics for Explainable Deep Learning

Recently, deep learning has been advancing the state of the art in artif...
research
11/11/2019

Explainable Artificial Intelligence (XAI) for 6G: Improving Trust between Human and Machine

As the 5th Generation (5G) mobile networks are bringing about global soc...
research
01/03/2023

Explainability and Robustness of Deep Visual Classification Models

In the computer vision community, Convolutional Neural Networks (CNNs), ...
research
05/04/2023

Neuro-symbolic model for cantilever beams damage detection

In the last decade, damage detection approaches swiftly changed from adv...

Please sign up or login with your details

Forgot password? Click here to reset