Interpretable and Trustworthy Deepfake Detection via Dynamic Prototypes
Authors: Loc Trinh, Michael Tsang, Sirisha Rambhatla, Yan Liu
Published: 2020-06-28 00:25:34+00:00
Comment: To appear in the 2021 IEEE Winter Conference on Applications of Computer Vision (WACV 21')
AI Summary
This paper introduces the Dynamic Prototype Network (DPNet), a novel human-centered approach for interpretable deepfake detection in face images within videos. DPNet utilizes dynamic prototypes to identify and explain temporal inconsistencies and artifacts, crucial for detecting sophisticated deepfakes. It achieves competitive predictive performance on various unseen datasets and enhances trustworthiness by formulating temporal logic specifications based on learned dynamics.
Abstract
In this paper we propose a novel human-centered approach for detecting forgery in face images, using dynamic prototypes as a form of visual explanations. Currently, most state-of-the-art deepfake detections are based on black-box models that process videos frame-by-frame for inference, and few closely examine their temporal inconsistencies. However, the existence of such temporal artifacts within deepfake videos is key in detecting and explaining deepfakes to a supervising human. To this end, we propose Dynamic Prototype Network (DPNet) -- an interpretable and effective solution that utilizes dynamic representations (i.e., prototypes) to explain deepfake temporal artifacts. Extensive experimental results show that DPNet achieves competitive predictive performance, even on unseen testing datasets such as Google's DeepFakeDetection, DeeperForensics, and Celeb-DF, while providing easy referential explanations of deepfake dynamics. On top of DPNet's prototypical framework, we further formulate temporal logic specifications based on these dynamics to check our model's compliance to desired temporal behaviors, hence providing trustworthiness for such critical detection systems.