IPGuard: Protecting the Intellectual Property of Deep Neural Networks via Fingerprinting the Classification Boundary

10/28/2019
by   Xiaoyu Cao, et al.
0

A deep neural network (DNN) classifier represents a model owner's intellectual property as training a DNN classifier often requires lots of resource. Watermarking was recently proposed to protect the intellectual property of DNN classifiers. However, watermarking suffers from a key limitation: it sacrifices the utility/accuracy of the model owner's classifier because it tampers the classifier's training or fine-tuning process. In this work, we propose IPGuard, the first method to protect intellectual property of DNN classifiers that provably incurs no accuracy loss for the classifiers. Our key observation is that a DNN classifier can be uniquely represented by its classification boundary. Based on this observation, IPGuard extracts some data points near the classification boundary of the model owner's classifier and uses them to fingerprint the classifier. A DNN classifier is said to be a pirated version of the model owner's classifier if they predict the same labels for most fingerprinting data points. IPGuard is qualitatively different from watermarking. Specifically, IPGuard extracts fingerprinting data points near the classification boundary of a classifier that is already trained, while watermarking embeds watermarks into a classifier during its training or fine-tuning process. We extensively evaluate IPGuard on CIFAR-10, CIFAR-100, and ImageNet datasets. Our results show that IPGuard can robustly identify post-processed versions of the model owner's classifier as pirated versions of the classifier, and IPGuard can identify classifiers, which are not the model owner's classifier nor its post-processed versions, as non-pirated versions of the classifier.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/08/2017

On Classification of Distorted Images with Deep Convolutional Neural Networks

Image blur and image noise are common distortions during image acquisiti...
research
08/01/2022

Backdoor Watermarking Deep Learning Classification Models With Deep Fidelity

Backdoor Watermarking is a promising paradigm to protect the copyright o...
research
08/16/2022

Neural network fragile watermarking with no model performance degradation

Deep neural networks are vulnerable to malicious fine-tuning attacks suc...
research
10/26/2020

Robust and Verifiable Information Embedding Attacks to Deep Neural Networks via Error-Correcting Codes

In the era of deep learning, a user often leverages a third-party machin...
research
02/13/2023

How Accurate Does It Feel? – Human Perception of Different Types of Classification Mistakes

Supervised machine learning utilizes large datasets, often with ground t...
research
03/07/2021

T-Miner: A Generative Approach to Defend Against Trojan Attacks on DNN-based Text Classification

Deep Neural Network (DNN) classifiers are known to be vulnerable to Troj...
research
02/10/2020

Stability for the Training of Deep Neural Networks and Other Classifiers

We examine the stability of loss-minimizing training processes that are ...

Please sign up or login with your details

Forgot password? Click here to reset