Probabilistic Verification of Voice Anti-Spoofing Models
Authors: Evgeny Kushnir, Alexandr Kozodaev, Dmitrii Korzh, Mikhail Pautov, Oleg Kiriukhin, Oleg Y. Rogov
Published: 2026-03-11 12:40:24+00:00
Comment: The paper was submitted for review to Interspeech 2026
AI Summary
The paper introduces PV-VASM, a probabilistic framework designed to verify the robustness of voice anti-spoofing models (VASMs) against malicious speech synthesis and input perturbations. It estimates the probability of misclassification under text-to-speech (TTS), voice cloning (VC), and parametric signal transformations, offering model-agnostic robustness guarantees. PV-VASM derives a theoretical upper bound on the error probability and is validated as a practical verification tool across diverse experimental settings.
Abstract
Recent advances in generative models have amplified the risk of malicious misuse of speech synthesis technologies, enabling adversaries to impersonate target speakers and access sensitive resources. Although speech deepfake detection has progressed rapidly, most existing countermeasures lack formal robustness guarantees or fail to generalize to unseen generation techniques. We propose PV-VASM, a probabilistic framework for verifying the robustness of voice anti-spoofing models (VASMs). PV-VASM estimates the probability of misclassification under text-to-speech (TTS), voice cloning (VC), and parametric signal transformations. The approach is model-agnostic and enables robustness verification against unseen speech synthesis techniques and input perturbations. We derive a theoretical upper bound on the error probability and validate the method across diverse experimental settings, demonstrating its effectiveness as a practical robustness verification tool.