DeepAI AI Chat
Log In Sign Up

Fortify Machine Learning Production Systems: Detect and Classify Adversarial Attacks

by   Matthew Ciolino, et al.

Production machine learning systems are consistently under attack by adversarial actors. Various deep learning models must be capable of accurately detecting fake or adversarial input while maintaining speed. In this work, we propose one piece of the production protection system: detecting an incoming adversarial attack and its characteristics. Detecting types of adversarial attacks has two primary effects: the underlying model can be trained in a structured manner to be robust from those attacks and the attacks can be potentially filtered out in realtime before causing any downstream damage. The adversarial image classification space is explored for models commonly used in transfer learning.


page 3

page 4


Adversarial attacks on an optical neural network

Adversarial attacks have been extensively investigated for machine learn...

Headless Horseman: Adversarial Attacks on Transfer Learning Models

Transfer learning facilitates the training of task-specific classifiers ...

A Modified Drake Equation for Assessing Adversarial Risk to Machine Learning Models

Each machine learning model deployed into production has a risk of adver...

Should Adversarial Attacks Use Pixel p-Norm?

Adversarial attacks aim to confound machine learning systems, while rema...

Btech thesis report on adversarial attack detection and purification of adverserially attacked images

This is Btech thesis report on detection and purification of adverserial...

Node Copying for Protection Against Graph Neural Network Topology Attacks

Adversarial attacks can affect the performance of existing deep learning...

Defending Against Adversarial Machine Learning

An Adversarial System to attack and an Authorship Attribution System (AA...