Adversarial Attacks Against Uncertainty Quantification

09/19/2023
by   Emanuele Ledda, et al.
0

Machine-learning models can be fooled by adversarial examples, i.e., carefully-crafted input perturbations that force models to output wrong predictions. While uncertainty quantification has been recently proposed to detect adversarial inputs, under the assumption that such attacks exhibit a higher prediction uncertainty than pristine data, it has been shown that adaptive attacks specifically aimed at reducing also the uncertainty estimate can easily bypass this defense mechanism. In this work, we focus on a different adversarial scenario in which the attacker is still interested in manipulating the uncertainty estimate, but regardless of the correctness of the prediction; in particular, the goal is to undermine the use of machine-learning models when their outputs are consumed by a downstream module or by a human operator. Following such direction, we: (i) design a threat model for attacks targeting uncertainty quantification; (ii) devise different attack strategies on conceptually different UQ techniques spanning for both classification and semantic segmentation problems; (iii) conduct a first complete and extensive analysis to compare the differences between some of the most employed UQ approaches under attack. Our extensive experimental analysis shows that our attacks are more effective in manipulating uncertainty quantification measures than attacks aimed to also induce misclassifications.

READ FULL TEXT
research
05/24/2020

Detecting Adversarial Examples for Speech Recognition via Uncertainty Quantification

Machine learning systems and also, specifically, automatic speech recogn...
research
04/10/2023

Gradient-based Uncertainty Attribution for Explainable Bayesian Deep Learning

Predictions made by deep learning models are prone to data perturbations...
research
02/08/2021

Exploiting epistemic uncertainty of the deep learning models to generate adversarial samples

Deep neural network architectures are considered to be robust to random ...
research
03/22/2021

ExAD: An Ensemble Approach for Explanation-based Adversarial Detection

Recent research has shown Deep Neural Networks (DNNs) to be vulnerable t...
research
02/02/2023

Randomized prior wavelet neural operator for uncertainty quantification

In this paper, we propose a novel data-driven operator learning framewor...
research
07/11/2022

DAUX: a Density-based Approach for Uncertainty eXplanations

Uncertainty quantification (UQ) is essential for creating trustworthy ma...
research
04/13/2023

Neural State-Space Models: Empirical Evaluation of Uncertainty Quantification

Effective quantification of uncertainty is an essential and still missin...

Please sign up or login with your details

Forgot password? Click here to reset