Adversarial Perturbations Fool Deepfake Detectors

03/24/2020
by   Apurva Gandhi, et al.
1

This work uses adversarial perturbations to enhance deepfake images and fool common deepfake detectors. We created adversarial perturbations using the Fast Gradient Sign Method and the Carlini and Wagner L2 norm attack in both blackbox and whitebox settings. Detectors achieved over 95 deepfakes, but less than 27 two improvements to deepfake detectors: (i) Lipschitz regularization, and (ii) Deep Image Prior (DIP). Lipschitz regularization constrains the gradient of the detector with respect to the input in order to increase robustness to input perturbations. The DIP defense removes perturbations using generative convolutional neural networks in an unsupervised manner. Regularization improved the detection of perturbed deepfakes on average, including a 10 accuracy boost in the blackbox case. The DIP defense achieved 95 perturbed deepfakes that fooled the original detector, while retaining 98 accuracy in other cases on a 100 image subsample.

READ FULL TEXT

page 1

page 2

page 6

research
02/14/2017

On Detecting Adversarial Perturbations

Machine learning and deep learning in particular has advanced tremendous...
research
08/18/2023

Attacking logo-based phishing website detectors with adversarial perturbations

Recent times have witnessed the rise of anti-phishing schemes powered by...
research
10/27/2019

EdgeFool: An Adversarial Image Enhancement Filter

Adversarial examples are intentionally perturbed images that mislead cla...
research
09/16/2018

Attacking Object Detectors via Imperceptible Patches on Background

Deep neural networks have been proven vulnerable against adversarial per...
research
07/11/2019

Why Blocking Targeted Adversarial Perturbations Impairs the Ability to Learn

Despite their accuracy, neural network-based classifiers are still prone...
research
02/10/2021

Towards Certifying ℓ_∞ Robustness using Neural Networks with ℓ_∞-dist Neurons

It is well-known that standard neural networks, even with a high classif...
research
08/20/2022

Evaluating Out-of-Distribution Detectors Through Adversarial Generation of Outliers

A reliable evaluation method is essential for building a robust out-of-d...

Please sign up or login with your details

Forgot password? Click here to reset