Exposing the Deception: Uncovering More Forgery Clues for Deepfake Detection

Authors: Zhongjie Ba, Qingyu Liu, Zhenguang Liu, Shuang Wu, Feng Lin, Li Lu, Kui Ren

Published: 2024-03-04 07:28:23+00:00

Comment: AAAI2024

AI Summary

This paper addresses challenges in deepfake detection, such as overfitting to local forgery clues and lack of theoretical constraints, which lead to unsatisfactory accuracy and limited generalizability. The authors propose a novel framework that captures broader forgery clues by extracting and fusing multiple non-overlapping local representations, guided by Local and Global Information Losses derived from information bottleneck theory. This approach ensures orthogonality of local features and eliminates task-irrelevant information, achieving state-of-the-art performance on five benchmark datasets.

Abstract

Deepfake technology has given rise to a spectrum of novel and compelling applications. Unfortunately, the widespread proliferation of high-fidelity fake videos has led to pervasive confusion and deception, shattering our faith that seeing is believing. One aspect that has been overlooked so far is that current deepfake detection approaches may easily fall into the trap of overfitting, focusing only on forgery clues within one or a few local regions. Moreover, existing works heavily rely on neural networks to extract forgery features, lacking theoretical constraints guaranteeing that sufficient forgery clues are extracted and superfluous features are eliminated. These deficiencies culminate in unsatisfactory accuracy and limited generalizability in real-life scenarios. In this paper, we try to tackle these challenges through three designs: (1) We present a novel framework to capture broader forgery clues by extracting multiple non-overlapping local representations and fusing them into a global semantic-rich feature. (2) Based on the information bottleneck theory, we derive Local Information Loss to guarantee the orthogonality of local representations while preserving comprehensive task-relevant information. (3) Further, to fuse the local representations and remove task-irrelevant information, we arrive at a Global Information Loss through the theoretical analysis of mutual information. Empirically, our method achieves state-of-the-art performance on five benchmark datasets.Our code is available at \\url{https://github.com/QingyuLiu/Exposing-the-Deception}, hoping to inspire researchers.


Key findings
The method achieved state-of-the-art performance across five benchmark datasets for both in-dataset and cross-dataset deepfake detection settings, demonstrating improved generalizability to unseen data. It successfully uncovers broader and disentangled forgery clues by focusing on multiple non-overlapping facial regions, such as the nose, cheek, forehead, and mouth, which contrasts with existing methods that often fixate on limited local areas. Ablation studies confirmed the critical contribution of both Local and Global Information Losses to the model's performance and robustness.
Approach
The proposed method captures broader forgery clues by extracting multiple non-overlapping local representations and fusing them into a global semantic-rich feature. It uses a Local Information Loss, based on information bottleneck theory, to guarantee the orthogonality of local representations while preserving comprehensive task-relevant information. A Global Information Loss, derived from mutual information analysis, further fuses these local representations and removes task-irrelevant information for robust classification.
Datasets
FaceForensics++ (FF++), Celeb-DF-V1 (CD1), Celeb-DF-V2 (CD2), DeepFake Detection Challenge (DFDC), DFDC-Preview (DFDC-P)
Model(s)
UNKNOWN
Author countries
China, Singapore