Topological Detection of Trojaned Neural Networks

06/11/2021
by   Songzhu Zheng, et al.
14

Deep neural networks are known to have security issues. One particular threat is the Trojan attack. It occurs when the attackers stealthily manipulate the model's behavior through Trojaned training samples, which can later be exploited. Guided by basic neuroscientific principles we discover subtle – yet critical – structural deviation characterizing Trojaned models. In our analysis we use topological tools. They allow us to model high-order dependencies in the networks, robustly compare different networks, and localize structural abnormalities. One interesting observation is that Trojaned models develop short-cuts from input to output layers. Inspired by these observations, we devise a strategy for robust detection of Trojaned models. Compared to standard baselines it displays better performance on multiple benchmarks.

READ FULL TEXT
research
06/08/2021

Provably Robust Detection of Out-of-distribution Data (almost) for free

When applying machine learning in safety-critical systems, a reliable as...
research
10/25/2022

Deep Neural Networks as the Semi-classical Limit of Topological Quantum Neural Networks: The problem of generalisation

Deep Neural Networks miss a principled model of their operation. A novel...
research
05/26/2018

Deep Learning Topological Invariants of Band Insulators

In this work we design and train deep neural networks to predict topolog...
research
12/04/2018

On incompressible high order networks

This work presents a theoretical investigation of incompressible high or...
research
06/15/2022

Architectural Backdoors in Neural Networks

Machine learning is vulnerable to adversarial manipulation. Previous lit...
research
01/21/2021

TDA-Net: Fusion of Persistent Homology and Deep Learning Features for COVID-19 Detection in Chest X-Ray Images

Topological Data Analysis (TDA) has emerged recently as a robust tool to...
research
02/15/2018

Putting a bug in ML: The moth olfactory network learns to read MNIST

We seek to (i) characterize the learning architectures exploited in biol...

Please sign up or login with your details

Forgot password? Click here to reset