Spatial-Temporal Frequency Forgery Clue for Video Forgery Detection in VIS and NIR Scenario

by   Yukai Wang, et al.

In recent years, with the rapid development of face editing and generation, more and more fake videos are circulating on social media, which has caused extreme public concerns. Existing face forgery detection methods based on frequency domain find that the GAN forged images have obvious grid-like visual artifacts in the frequency spectrum compared to the real images. But for synthesized videos, these methods only confine to single frame and pay little attention to the most discriminative part and temporal frequency clue among different frames. To take full advantage of the rich information in video sequences, this paper performs video forgery detection on both spatial and temporal frequency domains and proposes a Discrete Cosine Transform-based Forgery Clue Augmentation Network (FCAN-DCT) to achieve a more comprehensive spatial-temporal feature representation. FCAN-DCT consists of a backbone network and two branches: Compact Feature Extraction (CFE) module and Frequency Temporal Attention (FTA) module. We conduct thorough experimental assessments on two visible light (VIS) based datasets WildDeepfake and Celeb-DF (v2), and our self-built video forgery dataset DeepfakeNIR, which is the first video forgery dataset on near-infrared modality. The experimental results demonstrate the effectiveness of our method on detecting forgery videos in both VIS and NIR scenarios.


page 3

page 6


Spatiotemporal Inconsistency Learning for DeepFake Video Detection

The rapid development of facial manipulation techniques has aroused publ...

MD-CSDNetwork: Multi-Domain Cross Stitched Network for Deepfake Detection

The rapid progress in the ease of creating and spreading ultra-realistic...

FOCAL: A Forgery Localization Framework based on Video Coding Self-Consistency

Forgery operations on video contents are nowadays within the reach of an...

Detection of Deepfake Videos Using Long Distance Attention

With the rapid progress of deepfake techniques in recent years, facial v...

Dynamic texture analysis for detecting fake faces in video sequences

The creation of manipulated multimedia content involving human character...

I2V-GAN: Unpaired Infrared-to-Visible Video Translation

Human vision is often adversely affected by complex environmental factor...

DeepRhythm: Exposing DeepFakes with Attentional Visual Heartbeat Rhythms

As the GAN-based face image and video generation techniques, widely know...

Please sign up or login with your details

Forgot password? Click here to reset