Entangled Watermarks as a Defense against Model Extraction

02/27/2020
by   Hengrui Jia, et al.
9

Machine learning involves expensive data collection and training procedures. Model owners may be concerned that valuable intellectual property can be leaked if adversaries mount model extraction attacks. Because it is difficult to defend against model extraction without sacrificing significant prediction accuracy, watermarking leverages unused model capacity to have the model overfit to outlier input-output pairs, which are not sampled from the task distribution and are only known to the defender. The defender then demonstrates knowledge of the input-output pairs to claim ownership of the model at inference. The effectiveness of watermarks remains limited because they are distinct from the task distribution and can thus be easily removed through compression or other forms of knowledge transfer. We introduce Entangled Watermarking Embeddings (EWE). Our approach encourages the model to learn common features for classifying data that is sampled from the task distribution, but also data that encodes watermarks. An adversary attempting to remove watermarks that are entangled with legitimate data is also forced to sacrifice performance on legitimate data. Experiments on MNIST, Fashion-MNIST, and Google Speech Commands validate that the defender can claim model ownership with 95 copy, at a modest cost of 1

READ FULL TEXT

page 13

page 16

page 17

research
04/21/2021

Dataset Inference: Ownership Resolution in Machine Learning

With increasingly more data and computation involved in their training, ...
research
06/20/2023

FDINet: Protecting against DNN Model Extraction via Feature Distortion Index

Machine Learning as a Service (MLaaS) platforms have gained popularity d...
research
09/03/2019

High-Fidelity Extraction of Neural Network Models

Model extraction allows an adversary to steal a copy of a remotely deplo...
research
08/03/2017

Detection of Abnormal Input-Output Associations

We study a novel outlier detection problem that aims to identify abnorma...
research
12/07/2021

Defending against Model Stealing via Verifying Embedded External Features

Obtaining a well-trained model involves expensive data collection and tr...
research
10/11/2022

Generalization Analysis on Learning with a Concurrent Verifier

Machine learning technologies have been used in a wide range of practica...
research
06/13/2017

Getting deep recommenders fit: Bloom embeddings for sparse binary input/output networks

Recommendation algorithms that incorporate techniques from deep learning...

Please sign up or login with your details

Forgot password? Click here to reset