Open-World Deepfake Attribution via Confidence-Aware Asymmetric Learning

Authors: Haiyang Zheng, Nan Pu, Wenjing Li, Teng Long, Nicu Sebe, Zhun Zhong

Published: 2025-12-14 12:31:28+00:00

Comment: Accepted by AAAI2026

AI Summary

This paper proposes a Confidence-Aware Asymmetric Learning (CAL) framework for Open-World DeepFake Attribution (OW-DFA), which aims to attribute both known and unknown synthetic facial imagery. CAL addresses critical limitations of existing methods, specifically confidence skew leading to biased training and the unrealistic assumption of knowing the number of unknown forgery types. It achieves this through Confidence-Aware Consistency Regularization, Asymmetric Confidence Reinforcement, and a Dynamic Prototype Pruning strategy for automatic novel category estimation.

Abstract

The proliferation of synthetic facial imagery has intensified the need for robust Open-World DeepFake Attribution (OW-DFA), which aims to attribute both known and unknown forgeries using labeled data for known types and unlabeled data containing a mixture of known and novel types. However, existing OW-DFA methods face two critical limitations: 1) A confidence skew that leads to unreliable pseudo-labels for novel forgeries, resulting in biased training. 2) An unrealistic assumption that the number of unknown forgery types is known *a priori*. To address these challenges, we propose a Confidence-Aware Asymmetric Learning (CAL) framework, which adaptively balances model confidence across known and novel forgery types. CAL mainly consists of two components: Confidence-Aware Consistency Regularization (CCR) and Asymmetric Confidence Reinforcement (ACR). CCR mitigates pseudo-label bias by dynamically scaling sample losses based on normalized confidence, gradually shifting the training focus from high- to low-confidence samples. ACR complements this by separately calibrating confidence for known and novel classes through selective learning on high-confidence samples, guided by their confidence gap. Together, CCR and ACR form a mutually reinforcing loop that significantly improves the model's OW-DFA performance. Moreover, we introduce a Dynamic Prototype Pruning (DPP) strategy that automatically estimates the number of novel forgery types in a coarse-to-fine manner, removing the need for unrealistic prior assumptions and enhancing the scalability of our methods to real-world OW-DFA scenarios. Extensive experiments on the standard OW-DFA benchmark and a newly extended benchmark incorporating advanced manipulations demonstrate that CAL consistently outperforms previous methods, achieving new state-of-the-art performance on both known and novel forgery attribution.


Key findings
CAL consistently outperforms previous state-of-the-art methods on both known and novel forgery attribution across the OW-DFA-40 and original OW-DFA benchmarks. It achieves significant improvements in overall accuracy (e.g., 4.5% All ACC over CPL on OW-DFA-40) and notably enhances novel forgery attribution (e.g., 7.3% average Novel ACC over CPL). The Dynamic Prototype Pruning strategy successfully estimates the number of novel forgery types with high accuracy and robustness to initialization.
Approach
The proposed CAL framework incorporates a Frequency-Guided Feature Enhancement (FFE) module, Confidence-Aware Consistency Regularization (CCR) for adaptive loss scaling, and Asymmetric Confidence Reinforcement (ACR) for biased pseudo-label selection. Additionally, it features a Dynamic Prototype Pruning (DPP) strategy that dynamically merges low-usage prototypes to automatically estimate the number of novel forgery types in a coarse-to-fine manner.
Datasets
FaceForensics++, Celeb-DF (used to construct the OW-DFA-40 benchmark), original OW-DFA benchmark.
Model(s)
UNKNOWN
Author countries
Italy, China