Big Brother is Watching: Proactive Deepfake Detection via Learnable Hidden Face

Authors: Hongbo Li, Shangchao Yang, Ruiyang Xia, Lin Yuan, Xinbo Gao

Published: 2025-04-15 15:50:54+00:00

AI Summary

This paper proposes a novel proactive deepfake detection framework that embeds a learnable hidden face within a face image using an invertible steganography network. This hidden face acts as an indicator of malicious tampering, allowing for accurate deepfake detection even after common image processing or deepfake manipulations.

Abstract

As deepfake technologies continue to advance, passive detection methods struggle to generalize with various forgery manipulations and datasets. Proactive defense techniques have been actively studied with the primary aim of preventing deepfake operation effectively working. In this paper, we aim to bridge the gap between passive detection and proactive defense, and seek to solve the detection problem utilizing a proactive methodology. Inspired by several watermarking-based forensic methods, we explore a novel detection framework based on the concept of ``hiding a learnable face within a face''. Specifically, relying on a semi-fragile invertible steganography network, a secret template image is embedded into a host image imperceptibly, acting as an indicator monitoring for any malicious image forgery when being restored by the inverse steganography process. Instead of being manually specified, the secret template is optimized during training to resemble a neutral facial appearance, just like a ``big brother'' hidden in the image to be protected. By incorporating a self-blending mechanism and robustness learning strategy with a simulative transmission channel, a robust detector is built to accurately distinguish if the steganographic image is maliciously tampered or benignly processed. Finally, extensive experiments conducted on multiple datasets demonstrate the superiority of the proposed approach over competing passive and proactive detection methods.


Key findings
The proposed method outperforms existing passive and proactive deepfake detection methods across multiple datasets and forgery techniques. The approach maintains high image quality while exhibiting strong robustness to benign image processing. Ablation studies confirmed the importance of each component in the proposed framework.
Approach
The approach uses an invertible steganography network to embed a learnable secret template (a neutral face) into a face image. A deepfake detector analyzes the difference between the original and restored template after the image undergoes processing, identifying malicious manipulations based on discrepancies.
Datasets
FFHQ (for training), VGGFace2, and CelebA-HQ (for testing)
Model(s)
Invertible neural network (INN) for steganography, Patch discriminator for deepfake classification
Author countries
UNKNOWN