How to Label Resynthesized Audio: The Dual Role of Neural Audio Codecs in Audio Deepfake Detection
Authors: Yixuan Xiao, Florian Lux, Alejandro Pérez-González-de-Martos, Ngoc Thang Vu
Published: 2026-02-18 10:29:07+00:00
Comment: Accepted to ICASSP 2026
AI Summary
This study addresses the critical challenge of labeling resynthesized audio from neural audio codecs (NACs) in audio deepfake detection, given their dual role in compression and speech synthesis. The paper introduces a new, challenging dataset, CodecDeepfakeDetection (CDD), which extends ASVspoof 5. It thoroughly investigates how different labeling choices for codec-resynthesized audio (CoRS) affect deepfake detection performance and provides insights into optimal labeling strategies.
Abstract
Since Text-to-Speech systems typically don't produce waveforms directly, recent spoof detection studies use resynthesized waveforms from vocoders and neural audio codecs to simulate an attacker. Unlike vocoders, which are specifically designed for speech synthesis, neural audio codecs were originally developed for compressing audio for storage and transmission. However, their ability to discretize speech also sparked interest in language-modeling-based speech synthesis. Owing to this dual functionality, codec resynthesized data may be labeled as either bonafide or spoof. So far, very little research has addressed this issue. In this study, we present a challenging extension of the ASVspoof 5 dataset constructed for this purpose. We examine how different labeling choices affect detection performance and provide insights into labeling strategies.