Pixels Don't Lie (But Your Detector Might): Bootstrapping MLLM-as-a-Judge for Trustworthy Deepfake Detection and Reasoning Supervision

[CVPR 2026]

Mohamed bin Zayed University of Artificial Intelligence  ·  Monash University

📄 Abstract

Deepfake detection models increasingly generate natural language explanations to justify their predictions. Yet, while classification accuracy has improved, the reasoning behind these predictions is often ungrounded, hallucinated, or loosely connected to the actual visual evidence. Existing evaluation protocols focus primarily on detection accuracy and largely overlook reasoning fidelity, visual grounding, and interpretability.

We introduce DeepfakeJudge, a unified framework for scalable reasoning supervision and evaluation in deepfake detection. The framework integrates an out-of-distribution detection benchmark, a densely human-annotated reasoning dataset, and a bootstrapped generator-evaluator training pipeline to build a multimodal reasoning judge. The resulting models evaluate explanation quality directly from images and support both pointwise and pairwise assessment aligned with human judgment.

DeepfakeJudge establishes reasoning fidelity as a measurable and scalable dimension of trustworthy deepfake detection, showing that effective reasoning evaluators can be trained without requiring explicit ground-truth rationales for every instance.

🔍 Why DeepfakeJudge?

Current deepfake detectors can classify images, but their reasoning is often unreliable: hallucinated artifacts, ungrounded claims, and generic explanations remain widespread. Standard metrics such as BLEU and BERTScore do not capture these deficiencies. DeepfakeJudge addresses this gap by training compact vision-language models that assess reasoning quality directly from images.

Comparison of deepfake reasoning across different VLMs DeepfakeJudge scores vs traditional metrics
Figure 1: (Left) VLMs produce reasoning of widely varying quality for the same deepfake: some correctly identify artifacts, while others hallucinate non-existent manipulations. (Right) Traditional metrics (BLEU, BERTScore) fail to capture reasoning fidelity, whereas DeepfakeJudge scores correlate well with human assessments.

💡 Key Contributions

🎯

Out-of-Distribution Deepfake Benchmark

A challenging benchmark combining real images, text-to-image generations, and editing-based forgeries from modern pipelines such as Gemini, SeedDream, Flux-Kontext-Max, and Qwen-Edit, designed to evaluate both detection accuracy and reasoning generalization.

✍️

Human-Annotated Visual Reasoning Dataset

Densely annotated dataset linking textual explanations to localized visual evidence, covering artifact categories, bounding boxes, referring expressions, and structured gold reasoning rationales.

🔄

Bootstrapped Generator-Evaluator Pipeline

A scalable supervision framework that produces graded reasoning traces across five quality levels, iteratively refines misaligned samples using evaluator feedback, and paraphrases accepted outputs to introduce stylistic diversity.

⚖️

MLLM-as-a-Judge

Compact 3B and 7B vision-language models trained as reasoning evaluators, supporting both pointwise scoring (1 to 5) and pairwise comparison. The 7B model achieves 96.2% pairwise accuracy and 0.95 Pearson correlation, surpassing models over 30x larger.

📊 Dataset Construction

DeepfakeJudge introduces a multi-level dataset ecosystem: DFJ-Detect (2,000 images for OOD detection), DFJ-Reason (924 images with human-annotated reasoning), DFJ-Meta (41K+ bootstrapped training samples for judge training), and DFJ-Meta-Human (155 human-validated evaluation samples).

Data generation pipeline overview
Figure 3: Overview of the dataset construction pipeline. Real and synthetic images are curated, annotated by human experts, and expanded through bootstrapped reasoning generation to create graded supervision at scale.

🎯 DFJ-Detect

  • 1,000 real images (OpenImages-V7)
  • 500 T2I fakes (Gemini, SeedDream)
  • 500 edited fakes (Gemini, Flux-Kontext-Max, Qwen-Edit)

🧠 DFJ-Reason

  • 924 images with dense annotations
  • Artifact categories + bounding boxes
  • Referring expressions + gold rationales
  • Cohen's κ = 0.71 inter-annotator agreement

📷 Example Images from the Dataset

Real image example
REAL OpenImages-V7
T2I fake example
FAKE Gemini T2I
Edited image example
EDITED Flux-Kontext
Qwen edited example
EDITED Qwen-Edit

🔄 Bootstrapped Reasoning Supervision

To scale training beyond the human-annotated subset, we introduce a generator-evaluator loop. A generator produces reasoning traces at five intended quality levels (1 to 5). An evaluator checks alignment between predicted and intended ratings. Misaligned samples are refined using evaluator feedback until consistency is achieved. Accepted samples are then paraphrased to prevent stylistic memorization.

Bootstrapping generator-evaluator loop
Figure 4: The bootstrapped generator-evaluator supervision framework. Reasoning traces are generated at controlled quality levels, evaluated for alignment, iteratively refined, and diversified through paraphrasing.

📝 Example: Graded Reasoning from Bootstrapping

Example real image for bootstrapping
(Image label: Fake)
RatingReasoning
5 (Gold) The train in the mid-right appears unnaturally bright and overly smooth, lacking realistic surface textures that would normally show wear and detail. The entire upper-right train section is distorted, with shapes fused together and unclear details, making it hard to distinguish individual parts. The merging of elements creates an unnatural and confusing structure that does not align with typical train forms. Edges between the train and the background are uneven and lack depth, further weakening the image's realism. These texture and structural anomalies clearly indicate that the scene is artificially generated.
4 The train located on the right side is strikingly bright and shows a lack of detailed texture, giving it an overly smooth and synthetic appearance. In the upper-right section, the shapes seem to merge, complicating the identification of distinct train parts. The boundaries between the train and its background are somewhat uneven, further diminishing its natural look. These texture irregularities and the blending of forms strongly suggest that the image is AI-generated.
3 The train exhibits an unusually bright and smooth texture, with certain sections blending into one another in the upper right area. The distinct features of the train's structure are somewhat obscured, and the edges around it lack sharpness, indicating possible alterations. Overall, the image raises suspicion and does not appear entirely authentic, despite some visible details.
2 The train looks a little off because some parts seem merged and unclear, especially on the upper right. The colors also look a bit unnatural. However, the rest of the scene looks okay. It might be edited or just a low-quality photo. There are no strong clear signs, so I'm not completely sure.
1 The train looks normal and the tracks appear fine, with no visible issues. The colors and textures seem consistent with a real photo. The background and surrounding objects also look natural. Nothing stands out as fake or edited here, so this image is definitely real.
Table: Example of a fake image and corresponding degraded ratings produced by the bootstrapping process. Reasoning quality is systematically controlled from level 5 (fully accurate and grounded) to level 1 (hallucinated and misleading).

🔬 Methodology

Dataset Construction

Real and synthetic images are curated for an OOD detection benchmark. Fake images span both T2I generation and image-editing pipelines. A subset is densely annotated for reasoning supervision, linking textual explanations to spatial visual evidence.

Bootstrapped Reasoning Supervision

A generator model produces reasoning samples across five quality levels. An evaluator model assigns ratings and provides feedback. Misaligned samples are iteratively refined. Accepted samples are paraphrased to introduce stylistic diversity while preserving semantic structure.

DeepfakeJudge Training

Compact VLMs (3B and 7B) are fine-tuned using LoRA with a negative log-likelihood objective. In the pointwise setting, the model predicts a reasoning quality score (1 to 5) with a brief justification. In the pairwise setting, it selects the better-grounded reasoning between two candidates.

📈 Benchmark Results

🎯 Deepfake Detection (OOD)

Evaluation on DeepfakeJudge-Detect (2,000 images):

ModelReal F1Fake F1Overall Acc.
Gemini-2.5-Flash73.750.065.5
GPT-4o-mini70.235.859.3
Qwen-3-VL-235B78.668.474.5
Qwen-3-VL-235B-Thinking76.679.863.7
SIDA-13B57.034.548.1

🧠 Reasoning Evaluation

Evaluation on DeepfakeJudge-Reason:

ModelBLEU-3BERTScoreDFJ-3B Score
Gemini-2.5-Flash0.020.603.17
GPT-4o-mini0.010.352.83
Qwen-3-VL-30B0.030.623.31
Qwen-3-VL-235B0.010.603.59
SIDA0.010.582.32

📌 Pointwise Evaluation

DFJ-Meta

ModelRMSE ↓Pearson ↑
Gemini-2.51.090.83
GPT-4o-mini0.780.87
Qwen-3-VL-235B1.100.82
DeepfakeJudge-3B0.690.92
DeepfakeJudge-7B0.610.93

DFJ-Meta-Human

ModelRMSE ↓Pearson ↑
GPT-4o-mini0.810.86
Qwen-235B-Thinking0.950.86
DeepfakeJudge-7B0.500.95

⚖️ Pairwise Evaluation

Pairwise accuracy (% agreement with human preferences):

ModelDFJ-MetaDFJ-Meta-Human
Gemini-2.591.794.2
GPT-4o-mini90.389.8
Qwen-235B93.299.4
DeepfakeJudge-3B94.496.6
DeepfakeJudge-7B96.298.9

🦁 Model Zoo

All models are fine-tuned from Qwen2.5-VL-Instruct using LoRA and hosted on Hugging Face:

ModelTypeBaseLink
DeepfakeJudge-3B-Pointwise Pointwise Qwen2.5-VL-3B 🤗 Download
DeepfakeJudge-3B-Pairwise Pairwise Qwen2.5-VL-3B 🤗 Download
DeepfakeJudge-7B-Pointwise Pointwise Qwen2.5-VL-7B 🤗 Download
DeepfakeJudge-7B-Pairwise Pairwise Qwen2.5-VL-7B 🤗 Download

📝 Citation

If you find DeepfakeJudge useful in your research, please consider citing:

📢 arXiv paper coming soon. Citation will be updated upon release.