DFBench: Benchmarking Deepfake Image Detection Capability of Large Multimodal Models

Authors: Jiarui Wang, Huiyu Duan, Juntong Wang, Ziheng Jia, Woo Yi Yang, Xiaorong Zhu, Yu Zhao, Jiaying Qian, Yuke Xing, Guangtao Zhai, Xiongkuo Min

Published: 2025-06-03 15:45:41+00:00

AI Summary

This paper introduces DFBench, a large-scale deepfake image benchmark with diverse content (real, AI-edited, AI-generated) and state-of-the-art generative models. It proposes MoA-DF, a deepfake detection method using a combined probability strategy from multiple Large Multimodal Models (LMMs), achieving state-of-the-art performance.

Abstract

With the rapid advancement of generative models, the realism of AI-generated images has significantly improved, posing critical challenges for verifying digital content authenticity. Current deepfake detection methods often depend on datasets with limited generation models and content diversity that fail to keep pace with the evolving complexity and increasing realism of the AI-generated content. Large multimodal models (LMMs), widely adopted in various vision tasks, have demonstrated strong zero-shot capabilities, yet their potential in deepfake detection remains largely unexplored. To bridge this gap, we present textbf{DFBench}, a large-scale DeepFake Benchmark featuring (i) broad diversity, including 540,000 images across real, AI-edited, and AI-generated content, (ii) latest model, the fake images are generated by 12 state-of-the-art generation models, and (iii) bidirectional benchmarking and evaluating for both the detection accuracy of deepfake detectors and the evasion capability of generative models. Based on DFBench, we propose textbf{MoA-DF}, Mixture of Agents for DeepFake detection, leveraging a combined probability strategy from multiple LMMs. MoA-DF achieves state-of-the-art performance, further proving the effectiveness of leveraging LMMs for deepfake detection. Database and codes are publicly available at https://github.com/IntMeGroup/DFBench.


Key findings
MoA-DF achieves state-of-the-art performance on DFBench. The results highlight the increasing realism of AI-generated images and the limited generalization ability of existing deepfake detection methods. Large multimodal models show strong zero-shot generalization capabilities for deepfake detection.
Approach
MoA-DF solves the deepfake detection problem by combining the probabilistic outputs of multiple LMMs (Qwen2.5, InternVL2.5, InternVL3). This ensemble approach leverages the strengths of each LMM, mitigating individual model biases and improving robustness.
Datasets
DFBench (540,000 images: real images from LIVE, CSIQ, TID2013, KADID-10k, CLIVE, KonIQ-10k, and Flickr8k; AI-edited images from EPAIQA-15K; AI-generated images from 12 state-of-the-art models including PixArt-sigma, Playground, Kolors, SD3.5-Large, SD3-Medium, LaVi-Bridge, Kandinsky-3, Flux-schnell, Flux-dev, Janus, NOVA, and Infinity)
Model(s)
MoA-DF (an ensemble of Qwen2.5, InternVL2.5, and InternVL3 LMMs), various other LMMs (Llava-one-vision, DeepSeekVL, LLaVA-1.5, mPLUG-Owl3, CogAgent, InternLM-XComposer2.5, LLaVA-NeXT, Llama3.2-Vision, Qwen2-VL, Gemini1.5-pro, Grok2 Vision) and conventional deepfake detectors (CnnSpott, AntifakePrompt, Gram-Net, UnivFD, LGrad).
Author countries
China