When Deepfake Detection Meets Graph Neural Network:a Unified and Lightweight Learning Framework
Authors: Haoyu Liu, Chaoyu Gong, Mengke He, Jiate Li, Kai Han, Siqiang Luo
Published: 2025-08-07 15:55:13+00:00
AI Summary
This paper introduces SSTGNN, a lightweight deepfake detection framework using Spatial-Spectral-Temporal Graph Neural Networks. SSTGNN represents videos as structured graphs, enabling joint reasoning across spatial, temporal, and spectral domains, achieving superior performance with significantly fewer parameters than state-of-the-art models.
Abstract
The proliferation of generative video models has made detecting AI-generated and manipulated videos an urgent challenge. Existing detection approaches often fail to generalize across diverse manipulation types due to their reliance on isolated spatial, temporal, or spectral information, and typically require large models to perform well. This paper introduces SSTGNN, a lightweight Spatial-Spectral-Temporal Graph Neural Network framework that represents videos as structured graphs, enabling joint reasoning over spatial inconsistencies, temporal artifacts, and spectral distortions. SSTGNN incorporates learnable spectral filters and temporal differential modeling into a graph-based architecture, capturing subtle manipulation traces more effectively. Extensive experiments on diverse benchmark datasets demonstrate that SSTGNN not only achieves superior performance in both in-domain and cross-domain settings, but also offers strong robustness against unseen manipulations. Remarkably, SSTGNN accomplishes these results with up to 42.4$times$ fewer parameters than state-of-the-art models, making it highly lightweight and scalable for real-world deployment.