Relating Implicit Bias and Adversarial Attacks through Intrinsic Dimension

05/24/2023
by   Lorenzo Basile, et al.
0

Despite their impressive performance in classification, neural networks are known to be vulnerable to adversarial attacks. These attacks are small perturbations of the input data designed to fool the model. Naturally, a question arises regarding the potential connection between the architecture, settings, or properties of the model and the nature of the attack. In this work, we aim to shed light on this problem by focusing on the implicit bias of the neural network, which refers to its inherent inclination to favor specific patterns or outcomes. Specifically, we investigate one aspect of the implicit bias, which involves the essential Fourier frequencies required for accurate image classification. We conduct tests to assess the statistical relationship between these frequencies and those necessary for a successful attack. To delve into this relationship, we propose a new method that can uncover non-linear correlations between sets of coordinates, which, in our case, are the aforementioned frequencies. By exploiting the entanglement between intrinsic dimension and correlation, we provide empirical evidence that the network bias in Fourier space and the target frequencies of adversarial attacks are closely tied.

READ FULL TEXT

page 2

page 4

page 5

page 13

research
11/18/2019

A New Ensemble Adversarial Attack Powered by Long-term Gradient Memories

Deep neural networks are vulnerable to adversarial attacks....
research
11/24/2021

Thundernna: a white box adversarial attack

The existing work shows that the neural network trained by naive gradien...
research
03/13/2023

Can Adversarial Examples Be Parsed to Reveal Victim Model Information?

Numerous adversarial attack methods have been developed to generate impe...
research
11/24/2018

Attention, Please! Adversarial Defense via Attention Rectification and Preservation

This study provides a new understanding of the adversarial attack proble...
research
10/22/2022

Hindering Adversarial Attacks with Implicit Neural Representations

We introduce the Lossy Implicit Network Activation Coding (LINAC) defenc...
research
10/29/2020

Can the state of relevant neurons in a deep neural networks serve as indicators for detecting adversarial attacks?

We present a method for adversarial attack detection based on the inspec...
research
07/19/2023

Constructing Extreme Learning Machines with zero Spectral Bias

The phenomena of Spectral Bias, where the higher frequency components of...

Please sign up or login with your details

Forgot password? Click here to reset