CuRTAIL: ChaRacterizing and Thwarting AdversarIal deep Learning

09/08/2017
by   Bita Darvish Rouhani, et al.
0

This paper proposes CuRTAIL, an end-to-end computing framework for characterizing and thwarting adversarial space in the context of Deep Learning (DL). The framework protects deep neural networks against adversarial samples, which are perturbed inputs carefully crafted by malicious entities to mislead the underlying DL model. The precursor for the proposed methodology is a set of new quantitative metrics to assess the vulnerability of various deep learning architectures to adversarial samples. CuRTAIL formalizes the goal of preventing adversarial samples as a minimization of the space unexplored by the pertinent DL model that is characterized in CuRTAIL vulnerability analysis step. To thwart the adversarial machine learning attack, CuRTAIL introduces the concept of Modular Robust Redundancy (MRR) as a viable solution to achieve the formalized minimization objective. The MRR methodology explicitly characterizes the geometry of the input data and the DL model parameters. It then learns a set of complementary but disjoint models which maximally cover the unexplored subspaces of the target DL model, thus reducing the risk of integrity attacks. We extensively evaluate CuRTAIL performance against the state-of-the-art attack models including fast-sign-gradient, Jacobian Saliency Map Attack, and Deepfool. Proof-of-concept implementations for analyzing various data collections including MNIST, CIFAR10, and ImageNet corroborate CuRTAIL effectiveness to detect adversarial samples in different settings. The computations in each MRR module can be performed independently. As such, CuRTAIL detection algorithm can be completely parallelized among multiple hardware settings to achieve maximum throughput. We further provide an accompanying API to facilitate the adoption of the proposed framework for various applications.

READ FULL TEXT

page 8

page 10

research
04/10/2018

DeepMarks: A Digital Fingerprinting Framework for Deep Neural Networks

This paper proposes DeepMarks, a novel end-to-end framework for systemat...
research
07/31/2023

A Novel Deep Learning based Model to Defend Network Intrusion Detection System against Adversarial Attacks

Network Intrusion Detection System (NIDS) is an essential tool in securi...
research
06/14/2022

Downlink Power Allocation in Massive MIMO via Deep Learning: Adversarial Attacks and Training

The successful emergence of deep learning (DL) in wireless system applic...
research
09/13/2022

PINCH: An Adversarial Extraction Attack Framework for Deep Learning Models

Deep Learning (DL) models increasingly power a diversity of applications...
research
10/25/2022

A White-Box Adversarial Attack Against a Digital Twin

Recent research has shown that Machine Learning/Deep Learning (ML/DL) mo...
research
11/07/2022

Physics-Constrained Backdoor Attacks on Power System Fault Localization

The advances in deep learning (DL) techniques have the potential to deli...
research
04/02/2018

DeepSigns: A Generic Watermarking Framework for IP Protection of Deep Learning Models

This paper proposes DeepSigns, a novel end-to-end framework for systemat...

Please sign up or login with your details

Forgot password? Click here to reset