Sparse deepfake detection promotes better disentanglement
Authors: Antoine Teissier, Marie Tahon, Nicolas Dugué, Aghilas Sini
Published: 2025-10-07 09:03:39+00:00
AI Summary
This paper proposes a novel approach to enhance deepfake detection by introducing sparse latent representations in the AASIST architecture. By applying a TopK activation on the last hidden layer, the method improves detection performance, achieving an EER of 23.36% on ASVSpoof5 with 95% sparsity. Furthermore, it demonstrates that these sparse representations lead to better disentanglement of attack-related information in the latent space, thus promoting interpretability.
Abstract
Due to the rapid progress of speech synthesis, deepfake detection has become a major concern in the speech processing community. Because it is a critical task, systems must not only be efficient and robust, but also provide interpretable explanations. Among the different approaches for explainability, we focus on the interpretation of latent representations. In such paper, we focus on the last layer of embeddings of AASIST, a deepfake detection architecture. We use a TopK activation inspired by SAEs on this layer to obtain sparse representations which are used in the decision process. We demonstrate that sparse deepfake detection can improve detection performance, with an EER of 23.36% on ASVSpoof5 test set, with 95% of sparsity. We then show that these representations provide better disentanglement, using completeness and modularity metrics based on mutual information. Notably, some attacks are directly encoded in the latent space.