Reject Threshold Adaptation for Open-Set Model Attribution of Deepfake Audio

Authors: Xinrui Yan, Jiangyan Yi, Jianhua Tao, Yujie Chen, Hao Gu, Guanjun Li, Junzuo Zhou, Yong Ren, Tao Xu

Published: 2024-12-02 12:06:50+00:00

AI Summary

This paper proposes ReTA, a novel framework for open-set model attribution of deepfake audio, addressing the limitations of manually setting rejection thresholds in previous methods. ReTA adapts rejection thresholds for each class by learning reconstruction error distributions and employing Gaussian probability estimation, improving accuracy and data adaptability.

Abstract

Open environment oriented open set model attribution of deepfake audio is an emerging research topic, aiming to identify the generation models of deepfake audio. Most previous work requires manually setting a rejection threshold for unknown classes to compare with predicted probabilities. However, models often overfit training instances and generate overly confident predictions. Moreover, thresholds that effectively distinguish unknown categories in the current dataset may not be suitable for identifying known and unknown categories in another data distribution. To address the issues, we propose a novel framework for open set model attribution of deepfake audio with rejection threshold adaptation (ReTA). Specifically, the reconstruction error learning module trains by combining the representation of system fingerprints with labels corresponding to either the target class or a randomly chosen other class label. This process generates matching and non-matching reconstructed samples, establishing the reconstruction error distributions for each class and laying the foundation for the reject threshold calculation module. The reject threshold calculation module utilizes gaussian probability estimation to fit the distributions of matching and non-matching reconstruction errors. It then computes adaptive reject thresholds for all classes through probability minimization criteria. The experimental results demonstrate the effectiveness of ReTA in improving the open set model attributes of deepfake audio.


Key findings
ReTA outperforms existing methods (Softmax, OpenMax, CROSR) on the SFR dataset in terms of F1-score, total accuracy, and ID accuracy for both clean and compressed audio. The adaptive threshold approach improves the generalizability of open-set recognition for deepfake audio attribution.
Approach
ReTA uses a three-module architecture: system fingerprint recognition, reconstruction error learning, and adaptive reject threshold calculation. It learns reconstruction error distributions for known and simulated unknown classes, then calculates adaptive rejection thresholds using Gaussian probability estimation and a probability minimization criterion.
Datasets
SFR dataset (includes audio from Aispeech, Alibaba Cloud, Databaker, Sogou, Baidu Ai Cloud, Tencent, and iFLYTEK)
Model(s)
ResNet-18 based architecture with modifications for deepfake audio; uses a conditional sub-network and reconstruction decoder for error learning.
Author countries
China