FDINet: Protecting against DNN Model Extraction via Feature Distortion Index

06/20/2023
by   Hongwei Yao, et al.
0

Machine Learning as a Service (MLaaS) platforms have gained popularity due to their accessibility, cost-efficiency, scalability, and rapid development capabilities. However, recent research has highlighted the vulnerability of cloud-based models in MLaaS to model extraction attacks. In this paper, we introduce FDINET, a novel defense mechanism that leverages the feature distribution of deep neural network (DNN) models. Concretely, by analyzing the feature distribution from the adversary's queries, we reveal that the feature distribution of these queries deviates from that of the model's training set. Based on this key observation, we propose Feature Distortion Index (FDI), a metric designed to quantitatively measure the feature distribution deviation of received queries. The proposed FDINET utilizes FDI to train a binary detector and exploits FDI similarity to identify colluding adversaries from distributed extraction attacks. We conduct extensive experiments to evaluate FDINET against six state-of-the-art extraction attacks on four benchmark datasets and four popular model architectures. Empirical results demonstrate the following findings FDINET proves to be highly effective in detecting model extraction, achieving a 100 efficient, using just 50 queries to raise an extraction alarm with an average confidence of 96.08 colluding adversaries with an accuracy exceeding 91 demonstrates the ability to detect two types of adaptive attacks.

READ FULL TEXT
research
05/07/2018

PRADA: Protecting against DNN Model Stealing Attacks

As machine learning (ML) applications become increasingly prevalent, pro...
research
11/08/2021

DeepSteal: Advanced Model Extractions Leveraging Efficient Weight Stealing in Memories

Recent advancements of Deep Neural Networks (DNNs) have seen widespread ...
research
02/27/2020

Entangled Watermarks as a Defense against Model Extraction

Machine learning involves expensive data collection and training procedu...
research
04/14/2023

Interpretability is a Kind of Safety: An Interpreter-based Ensemble for Adversary Defense

While having achieved great success in rich real-life applications, deep...
research
11/20/2017

Model Extraction Warning in MLaaS Paradigm

Cloud vendors are increasingly offering machine learning services as par...
research
06/21/2021

Hardness of Samples Is All You Need: Protecting Deep Learning Models Using Hardness of Samples

Several recent studies have shown that Deep Neural Network (DNN)-based c...
research
02/21/2022

HoneyModels: Machine Learning Honeypots

Machine Learning is becoming a pivotal aspect of many systems today, off...

Please sign up or login with your details

Forgot password? Click here to reset