ConvNets and ImageNet Beyond Accuracy: Explanations, Bias Detection, Adversarial Examples and Model Criticism

11/30/2017
by   Pierre Stock, et al.
0

ConvNets and Imagenet have driven the recent success of deep learning for image classification. However, the marked slowdown in performance improvement, the recent studies on the lack of robustness of neural networks to adversarial examples and their tendency to exhibit undesirable biases (e.g racial biases) questioned the reliability and the sustained development of these methods. This work investigates these questions from the perspective of the end-user by using human subject studies and explanations. We experimentally demonstrate that the accuracy and robustness of ConvNets measured on Imagenet are underestimated. We show that explanations can mitigate the impact of misclassified adversarial examples from the perspective of the end-user and we introduce a novel tool for uncovering the undesirable biases learned by a model. These contributions also show that explanations are a promising tool for improving our understanding of ConvNets' predictions and for designing more reliable models

READ FULL TEXT

page 1

page 5

page 6

page 7

page 8

research
08/05/2018

Is Robustness the Cost of Accuracy? -- A Comprehensive Study on the Robustness of 18 Deep Image Classification Models

The prediction accuracy has been the long-lasting and sole standard for ...
research
02/25/2020

Gödel's Sentence Is An Adversarial Example But Unsolvable

In recent years, different types of adversarial examples from different ...
research
09/11/2020

Counterfactual Explanations Adversarial Examples – Common Grounds, Essential Differences, and Potential Transfers

It is well known that adversarial examples and counterfactual explanatio...
research
07/05/2021

When and How to Fool Explainable Models (and Humans) with Adversarial Examples

Reliable deployment of machine learning models such as neural networks c...
research
10/19/2020

Optimism in the Face of Adversity: Understanding and Improving Deep Learning through Adversarial Robustness

Driven by massive amounts of data and important advances in computationa...
research
02/23/2021

Rethinking Natural Adversarial Examples for Classification Models

Recently, it was found that many real-world examples without intentional...
research
06/07/2019

Reliable Classification Explanations via Adversarial Attacks on Robust Networks

Neural Networks (NNs) have been found vulnerable to a class of impercept...

Please sign up or login with your details

Forgot password? Click here to reset