Seeds Don't Lie: An Adaptive Watermarking Framework for Computer Vision Models

11/24/2022
by   Jacob Shams, et al.
0

In recent years, various watermarking methods were suggested to detect computer vision models obtained illegitimately from their owners, however they fail to demonstrate satisfactory robustness against model extraction attacks. In this paper, we present an adaptive framework to watermark a protected model, leveraging the unique behavior present in the model due to a unique random seed initialized during the model training. This watermark is used to detect extracted models, which have the same unique behavior, indicating an unauthorized usage of the protected model's intellectual property (IP). First, we show how an initial seed for random number generation as part of model training produces distinct characteristics in the model's decision boundaries, which are inherited by extracted models and present in their decision boundaries, but aren't present in non-extracted models trained on the same data-set with a different seed. Based on our findings, we suggest the Robust Adaptive Watermarking (RAW) Framework, which utilizes the unique behavior present in the protected and extracted models to generate a watermark key-set and verification model. We show that the framework is robust to (1) unseen model extraction attacks, and (2) extracted models which undergo a blurring method (e.g., weight pruning). We evaluate the framework's robustness against a naive attacker (unaware that the model is watermarked), and an informed attacker (who employs blurring strategies to remove watermarked behavior from an extracted model), and achieve outstanding (i.e., >0.9) AUC values. Finally, we show that the framework is robust to model extraction attacks with different structure and/or architecture than the protected model.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/06/2020

Training DNN Model with Secret Key for Model Protection

In this paper, we propose a model protection method by using block-wise ...
research
05/31/2021

A Protection Method of Trained CNN Model with Secret Key from Unauthorized Access

In this paper, we propose a novel method for protecting convolutional ne...
research
07/02/2023

Tools for Verifying Neural Models' Training Data

It is important that consumers and regulators can verify the provenance ...
research
02/06/2023

Protecting Language Generation Models via Invisible Watermarking

Language generation models have been an increasingly powerful enabler fo...
research
06/03/2019

DAWN: Dynamic Adversarial Watermarking of Neural Networks

Training machine learning (ML) models is expensive in terms of computati...
research
07/26/2022

Generative Extraction of Audio Classifiers for Speaker Identification

It is perhaps no longer surprising that machine learning models, especia...
research
10/07/2022

To tree or not to tree? Assessing the impact of smoothing the decision boundaries

When analyzing a dataset, it can be useful to assess how smooth the deci...

Please sign up or login with your details

Forgot password? Click here to reset