Who Made This? Fake Detection and Source Attribution with Diffusion Features
Authors: Simone Bonechi, Paolo Andreini, Barbara Toniella Corradini
Published: 2025-10-31 16:27:34+00:00
AI Summary
The paper introduces FRIDA, a lightweight and training-free framework that leverages internal latent activations (diffusion features) from a pre-trained Stable Diffusion Model (SDM) U-Net for synthetic image forensics. FRIDA achieves state-of-the-art cross-generator performance for deepfake detection using a k-NN classifier and enables accurate source generator attribution using a compact neural model. The results indicate that diffusion representations inherently encode robust, generator-specific patterns useful for detecting and attributing synthetic media.
Abstract
The rapid progress of generative diffusion models has enabled the creation of synthetic images that are increasingly difficult to distinguish from real ones, raising concerns about authenticity, copyright, and misinformation. Existing supervised detectors often struggle to generalize across unseen generators, requiring extensive labeled data and frequent retraining. We introduce FRIDA (Fake-image Recognition and source Identification via Diffusion-features Analysis), a lightweight framework that leverages internal activations from a pre-trained diffusion model for deepfake detection and source generator attribution. A k-nearest-neighbor classifier applied to diffusion features achieves state-of-the-art cross-generator performance without fine-tuning, while a compact neural model enables accurate source attribution. These results show that diffusion representations inherently encode generator-specific patterns, providing a simple and interpretable foundation for synthetic image forensics.