Prompt Tuning for Audio Deepfake Detection: Computationally Efficient Test-time Domain Adaptation with Limited Target Dataset

Authors: Hideyuki Oiso, Yuto Matsunaga, Kazuya Kakizaki, Taiki Miyagawa

Published: 2024-10-13 15:07:35+00:00

Comment: Accepted at Interspeech 2024. Hideyuki Oiso and Yuto Matsunaga contributed equally

AI Summary

This paper proposes a prompt tuning method for Audio Deepfake Detection (ADD) to address critical challenges in test-time domain adaptation, including source-target domain gaps, limited target dataset sizes, and high computational costs. The method operates in a plug-in style, seamlessly integrating with state-of-the-art transformer models to enhance accuracy on target data. By introducing a small number of trainable parameters, it prevents overfitting on small datasets and maintains computational efficiency.

Abstract

We study test-time domain adaptation for audio deepfake detection (ADD), addressing three challenges: (i) source-target domain gaps, (ii) limited target dataset size, and (iii) high computational costs. We propose an ADD method using prompt tuning in a plug-in style. It bridges domain gaps by integrating it seamlessly with state-of-the-art transformer models and/or with other fine-tuning methods, boosting their performance on target data (challenge (i)). In addition, our method can fit small target datasets because it does not require a large number of extra parameters (challenge (ii)). This feature also contributes to computational efficiency, countering the high computational costs typically associated with large-scale pre-trained models in ADD (challenge (iii)). We conclude that prompt tuning for ADD under domain gaps presents a promising avenue for enhancing accuracy with minimal target data and negligible extra computational burden.


Key findings
Prompt tuning consistently improved or maintained Equal Error Rates (EERs) across various target domains for two SOTA ADD models, even with extremely limited target data (as few as 10 samples), where full fine-tuning tends to overfit. The method demonstrated high computational efficiency, requiring minimal additional trainable parameters (e.g., 0.00161% for W2V) and showing rapid performance saturation with short prompt lengths (~10).
Approach
The authors propose a plug-in style prompt tuning method for test-time domain adaptation in Audio Deepfake Detection. Trainable prompt parameters are inserted into the intermediate feature vectors of pre-trained transformer-based models and fine-tuned on a small labeled target dataset. This approach allows for efficient adaptation by minimizing additional trainable parameters, thus avoiding overfitting and reducing computational burden.
Datasets
ASVspoof 2019 LA, In-The-Wild, Hamburg Adult Bilingual LAnguage (HABLA), ASVspoof 2021 LA, Voice Conversion Challenge (VCC) 2020
Model(s)
wav2vec 2.0, AASIST, Whisper, MesoNet
Author countries
Japan