What is it about?
AI can now create pictures that look completely real. This is great for art and design, but it also means people could use these "deepfakes" to spread false information. How can we tell if a photo is genuine or made by an AI? Our research tackles this problem in two main ways: We built a better test. Previous tests for "fake image detectors" were too easy and didn't use the latest AI fakes. We created a huge new test, called DFBench, with over half a million images. It includes modern AI fakes, partially edited photos, and real photos with flaws to make detection much more challenging and realistic. We created a smarter detective. Instead of building a specialized tool, we found that the latest, most powerful AI models (the kind that can chat and reason) are surprisingly good at spotting fakes. We then developed a method called MoA-DF that combines the "opinions" of several of these smart AIs to create an even more reliable detective. In short: We've built the most challenging exam to date for spotting AI-generated images, and we've shown that using teams of general-purpose smart AIs is a very effective way to pass it.
Featured Image
Photo by Nahrizul Kadri on Unsplash
Why is it important?
AI image generators have become so advanced that their creations are often indistinguishable from real photographs to the human eye. This explosion in capability is happening right now, creating an urgent need for equally advanced detection methods. Our work directly addresses this pressing, real-world problem that affects misinformation, digital trust, and media integrity.
Perspectives
Throughout this project, I was driven by the realization that we are fundamentally unprepared for the wave of synthetic media that is already upon us. Existing detection methods felt like trying to stop a modern cyberattack with an antique firewall—they were built for a different era. Seeing state-of-the-art detectors fail on images from the latest generative models was not just a research result; it was a warning. This led to our core insight: we must fight generality with generality. The new generation of AI that creates these stunning images is broad, creative, and adaptable. Our defenses must be the same. It was incredibly exciting to discover that large multimodal models (LMMs), which were not specifically designed for this task, possessed a latent, powerful ability to discern authenticity. It felt like we were uncovering a new, innate sense in these models. Developing the MoA-DF method was our way of formalizing this intuition. The "Mixture of Agents" approach is more than just an ensemble technique; it's a philosophical stance. In a world flooded with increasingly perfect forgeries, there may be no single "silver bullet." Instead, trust must emerge from a consensus of diverse, reasoning systems, much like it does in human societies. My hope is that DFBench will serve not just as a benchmark, but as a catalyst. I want it to push the entire field to aim higher, to stop chasing yesterday's fakes and start building the resilient, intelligent systems we need to secure our digital future. This work is our contribution to that vital effort.
Jiarui Wang
Read the Original
This page is a summary of: DFBench: Benchmarking Deepfake Image Detection Capability of Large Multimodal Models, October 2025, ACM (Association for Computing Machinery),
DOI: 10.1145/3746027.3758204.
You can read the full text:
Resources
Contributors
The following have contributed to this page







