Finite Gaussian Neurons: Defending against adversarial attacks by making neural networks say "I don't know"

06/13/2023
by   Felix Grezes, et al.
0

Since 2014, artificial neural networks have been known to be vulnerable to adversarial attacks, which can fool the network into producing wrong or nonsensical outputs by making humanly imperceptible alterations to inputs. While defenses against adversarial attacks have been proposed, they usually involve retraining a new neural network from scratch, a costly task. In this work, I introduce the Finite Gaussian Neuron (FGN), a novel neuron architecture for artificial neural networks. My works aims to: - easily convert existing models to Finite Gaussian Neuron architecture, - while preserving the existing model's behavior on real data, - and offering resistance against adversarial attacks. I show that converted and retrained Finite Gaussian Neural Networks (FGNN) always have lower confidence (i.e., are not overconfident) in their predictions over randomized and Fast Gradient Sign Method adversarial images when compared to classical neural networks, while maintaining high accuracy and confidence over real MNIST images. To further validate the capacity of Finite Gaussian Neurons to protect from adversarial attacks, I compare the behavior of FGNs to that of Bayesian Neural Networks against both randomized and adversarial images, and show how the behavior of the two architectures differs. Finally I show some limitations of the FGN models by testing them on the more complex SPEECHCOMMANDS task, against the stronger Carlini-Wagner and Projected Gradient Descent adversarial attacks.

READ FULL TEXT

page 14

page 16

page 25

page 27

page 30

page 32

page 36

page 38

research
04/18/2019

Gotta Catch 'Em All: Using Concealed Trapdoors to Detect Adversarial Attacks on Neural Networks

Deep neural networks are vulnerable to adversarial attacks. Numerous eff...
research
03/06/2023

Visual Analytics of Neuron Vulnerability to Adversarial Attacks on Convolutional Neural Networks

Adversarial attacks on a convolutional neural network (CNN) – injecting ...
research
07/04/2020

Relationship between manifold smoothness and adversarial vulnerability in deep learning with local errors

Artificial neural networks can achieve impressive performances, and even...
research
12/18/2019

Adversarial VC-dimension and Sample Complexity of Neural Networks

Adversarial attacks during the testing phase of neural networks pose a c...
research
11/11/2020

Adversarial images for the primate brain

Deep artificial neural networks have been proposed as a model of primate...
research
07/12/2021

SoftHebb: Bayesian inference in unsupervised Hebbian soft winner-take-all networks

State-of-the-art artificial neural networks (ANNs) require labelled data...
research
03/29/2023

A Tensor-based Convolutional Neural Network for Small Dataset Classification

Inspired by the ConvNets with structured hidden representations, we prop...

Please sign up or login with your details

Forgot password? Click here to reset