Model Extraction and Defenses on Generative Adversarial Networks

01/06/2021
by   Hailong Hu, et al.
14

Model extraction attacks aim to duplicate a machine learning model through query access to a target model. Early studies mainly focus on discriminative models. Despite the success, model extraction attacks against generative models are less well explored. In this paper, we systematically study the feasibility of model extraction attacks against generative adversarial networks (GANs). Specifically, we first define accuracy and fidelity on model extraction attacks against GANs. Then we study model extraction attacks against GANs from the perspective of accuracy extraction and fidelity extraction, according to the adversary's goals and background knowledge. We further conduct a case study where an adversary can transfer knowledge of the extracted model which steals a state-of-the-art GAN trained with more than 3 million images to new domains to broaden the scope of applications of model extraction attacks. Finally, we propose effective defense techniques to safeguard GANs, considering a trade-off between the utility and security of GAN models.

READ FULL TEXT

page 9

page 10

page 15

page 16

page 18

page 19

research
06/08/2023

Ownership Protection of Generative Adversarial Networks

Generative adversarial networks (GANs) have shown remarkable success in ...
research
10/06/2020

BAAAN: Backdoor Attacks Against Autoencoder and GAN-Based Machine Learning Models

The tremendous progress of autoencoders and generative adversarial netwo...
research
05/29/2023

NaturalFinger: Generating Natural Fingerprint with Generative Adversarial Networks

Deep neural network (DNN) models have become a critical asset of the mod...
research
06/13/2019

Improving Prediction Accuracy in Building Performance Models Using Generative Adversarial Networks (GANs)

Building performance discrepancies between building design and operation...
research
09/03/2019

High-Fidelity Extraction of Neural Network Models

Model extraction allows an adversary to steal a copy of a remotely deplo...
research
04/26/2021

Good Artists Copy, Great Artists Steal: Model Extraction Attacks Against Image Translation Generative Adversarial Networks

Machine learning models are typically made available to potential client...
research
01/24/2022

Attacks and Defenses for Free-Riders in Multi-Discriminator GAN

Generative Adversarial Networks (GANs) are increasingly adopted by the i...

Please sign up or login with your details

Forgot password? Click here to reset