Pixels Don't Lie (But Your Detector Might): Bootstrapping MLLM-as-a-Judge for Trustworthy Deepfake Detection and Reasoning Supervision
Authors: Kartik Kuckreja, Parul Gupta, Muhammad Haris Khan, Abhinav Dhall
Published: 2026-02-23 11:08:46+00:00
Comment: CVPR-2026, Code is available here: https://github.com/KjAeRsTuIsK/DeepfakeJudge
AI Summary
The paper introduces DeepfakeJudge, a framework for scalable reasoning supervision and evaluation in deepfake detection. It integrates an out-of-distribution benchmark, a human-annotated dataset with visual reasoning labels, and VLM-based evaluation models. DeepfakeJudge, optimized via a bootstrapped generator-evaluator process, achieves 96.2% accuracy and high human correlation in reasoning assessment, establishing reasoning fidelity as a quantifiable dimension for trustworthy deepfake detection.
Abstract
Deepfake detection models often generate natural-language explanations, yet their reasoning is frequently ungrounded in visual evidence, limiting reliability. Existing evaluations measure classification accuracy but overlook reasoning fidelity. We propose DeepfakeJudge, a framework for scalable reasoning supervision and evaluation, that integrates an out-of-distribution benchmark containing recent generative and editing forgeries, a human-annotated subset with visual reasoning labels, and a suite of evaluation models, that specialize in evaluating reasoning rationales without the need for explicit ground truth reasoning rationales. The Judge is optimized through a bootstrapped generator-evaluator process that scales human feedback into structured reasoning supervision and supports both pointwise and pairwise evaluation. On the proposed meta-evaluation benchmark, our reasoning-bootstrapped model achieves an accuracy of 96.2\\%, outperforming \\texttt{30x} larger baselines. The reasoning judge attains very high correlation with human ratings and 98.9\\% percent pairwise agreement on the human-annotated meta-evaluation subset. These results establish reasoning fidelity as a quantifiable dimension of deepfake detection and demonstrate scalable supervision for interpretable deepfake reasoning. Our user study shows that participants preferred the reasonings generated by our framework 70\\% of the time, in terms of faithfulness, groundedness, and usefulness, compared to those produced by other models and datasets. All of our datasets, models, and codebase are \\href{https://github.com/KjAeRsTuIsK/DeepfakeJudge}{open-sourced}.