Examining the Difference Among Transformers and CNNs with Explanation Methods

12/13/2022
by   Mingqi Jiang, et al.
0

We propose a methodology that systematically applies deep explanation algorithms on a dataset-wide basis, to compare different types of visual recognition backbones, such as convolutional networks (CNNs), global attention networks, and local attention networks. Examination of both qualitative visualizations and quantitative statistics across the dataset helps us to gain intuitions that are not just anecdotal, but are supported by the statistics computed on the entire dataset. Specifically, we propose two methods. The first one, sub-explanation counting, systematically searches for minimally-sufficient explanations of all images and count the amount of sub-explanations for each network. The second one, called cross-testing, computes salient regions using one network and then evaluates the performance by only showing these regions as an image to other networks. Through a combination of qualitative insights and quantitative statistics, we illustrate that 1) there are significant differences between the salient features of CNNs and attention models; 2) the occlusion-robustness in local attention models and global attention models may come from different decision-making mechanisms.

READ FULL TEXT

page 18

page 19

page 20

page 21

page 23

page 25

page 26

page 27

research
10/29/2021

ST-ABN: Visual Explanation Taking into Account Spatio-temporal Information for Video Recognition

It is difficult for people to interpret the decision-making in the infer...
research
03/27/2023

UFO: A unified method for controlling Understandability and Faithfulness Objectives in concept-based explanations for CNNs

Concept-based explanations for convolutional neural networks (CNNs) aim ...
research
06/25/2020

Explainable CNN-attention Networks (C-Attention Network) for Automated Detection of Alzheimer's Disease

In this work, we propose three explainable deep learning architectures t...
research
05/29/2021

EDDA: Explanation-driven Data Augmentation to Improve Model and Explanation Alignment

Recent years have seen the introduction of a range of methods for post-h...
research
12/10/2020

Debiased-CAM for bias-agnostic faithful visual explanations of deep convolutional networks

Class activation maps (CAMs) explain convolutional neural network predic...
research
09/14/2020

SCOUTER: Slot Attention-based Classifier for Explainable Image Recognition

Explainable artificial intelligence is gaining attention. However, most ...
research
05/18/2018

A Theoretical Explanation for Perplexing Behaviors of Backpropagation-based Visualizations

Backpropagation-based visualizations have been proposed to interpret con...

Please sign up or login with your details

Forgot password? Click here to reset