AI Images Are Everywhere. We’re All Deepfake Suspects Now.

Why AI image detectors are repeating the same disastrous mistakes as text detectors

January 6, 20269 min read
AI Images Are Everywhere. We’re All Deepfake Suspects Now.

Scroll through Twitter (or X, if you're pedantic) for five minutes, and you'll see AI-generated images created by Grok everywhere.

Browse through your Instagram feed and count how many images were touched by Midjourney or DALL-E.

Check your LinkedIn connections' profile pictures. Spot the telltale smoothness of AI enhancement.

The numbers are staggering: approximately 34 million new AI-generated images are created every day. The quality has improved exponentially. Gone are the days of obviously distorted hands and uncanny valley faces.

Modern AI image generators produce photos so realistic that even AI-savvy adults can only identify AI-generated images about half the time. The technology has become so sophisticated that 71% of images shared on social media globally are now AI-generated.

This isn't some distant future scenario. It's happening right now, and it's accelerating. With this explosion came the inevitable response: AI image detectors, positioned as the solution to our authenticity crisis.

Tech companies rushed to market with tools promising to separate the real from the synthetic. Investors poured money into detection startups. Platforms began implementing automated checks.

We're about to make the same mistakes all over again.

The AI Writing Detector Disaster We're Ignoring

Some of the most popular AI writing detectors.

Remember when AI writing detectors were going to save academic integrity? Universities rushed to implement tools like GPTZero and Turnitin, confident they could catch students using ChatGPT.

In October 2025, ABC News reported that Australian Catholic University had used AI detection technology to accuse approximately 6,000 students of academic misconduct in a single year. The problem? Many of those accusations were wrong.

A paramedic student found himself fighting to prove his innocence after being flagged for cheating on work he'd written himself. The burden of proof had shifted. Suddenly, students weren't presumed innocent until proven guilty. They were presumed guilty until they could somehow prove they hadn't used AI.

This isn't an isolated case. NPR reported in December 2025 that school districts across the United States are spending thousands of dollars on AI detection tools "despite research showing the technology is far from reliable."

At the University of Houston-Downtown, a computer science student named Leigh Burrell received a zero on an assignment worth 15% of her grade. Her professor believed she'd used ChatGPT. But The New York Times reviewed her Google Docs editing history and found she'd drafted and revised the paper over two days. She'd done the work. The detector got it wrong anyway.

Non-native English writers. Neurodivergent students. Anyone whose writing style happened to align with statistical patterns the detectors were trained to recognise. They all became suspects.

OpenAI itself quietly retired its AI text detection classifier after achieving just a 26% success rate. Think about that:

The company that created ChatGPT couldn't reliably detect its own AI's output.

Yet here we are, ready to deploy image detectors with the same blind faith.

Why AI Image Detectors Will Follow the Same Path

Photo by Nahrizul Kadri on Unsplash

The fundamental problem is identical: AI image generators are constantly improving, and as these platforms evolve, they create increasingly difficult images for detectors to flag. It's an arms race where the detection tools are perpetually playing catch-up.

Current AI image detectors perform no better than a coin toss in many cases. Researchers at Cornell University developed the Visual Counter Turing Test, a benchmark testing 17 leading AI image detection models against 166,000 images.

The result? An average accuracy of just 58%. That's 8% points better than random guessing.

Even more troubling: a study from Nexcess found that even AI-savvy adults could identify AI-generated images only about half the time.

If humans who understand AI struggle to distinguish real from synthetic, why are we confident that automated tools will do better?

The Paranoia Economy and Weaponisation of Doubt

Here are two things that worry me most.

First, the emergence of a paranoia economy where everyone becomes suspect.

Imagine you're a photographer who spent hours capturing the perfect wildlife shot. You submit it to a competition. The organisers run it through an AI detector. The algorithm flags it. Suddenly, you're defending your credibility, explaining your workflow, providing metadata, and essentially proving your innocence.

Or perhaps you're a graphic designer creating marketing materials. A client runs your work through a detector. False positive. Now you're justifying your artistic choices and professional expertise, all because an algorithm thinks your colour grading looks "too perfect" or your composition is "statistically improbable".

This isn't hypothetical fearmongering. We've already seen it happen with text detectors, where students had to prove they wrote their own essays.

Photo by Levi Arnold on Unsplash.webp

Second, the weaponisation of doubt.

In August 2024, Donald Trump looked at a photograph of thousands of people gathered at a Kamala Harris rally in Detroit. “There was nobody at the plane, and she ‘A.I.’d’ it,” he wrote on Truth Social. The crowd, he claimed, “DIDN’T EXIST!”

Trump's original post on his Truth Social account

BBC Verify confirmed the rally was real. The New York Times confirmed it. Dozens of attendees posted their own photos and videos from inside the crowd. News crews filmed the entire event.

None of that mattered. The accusation spread and sparked debate on social media.

An AI researcher told NPR:

"This is a photo of an event in one city on one day. I mean, what hope do we have to actually tackle complex problems in society if we can't agree on this?"

This is the endgame of unreliable detection. Not a world where we catch every fake, but a world where nothing can be proven real. Where the existence of AI images creates enough uncertainty to dismiss anything inconvenient.

When AI Image Detectors Actually Help

I'm not arguing that these tools have no place in our digital ecosystem. They do.

There are cases where it serves the public interest. But those cases share specific characteristics worth understanding

In May 2023, an image showing black smoke billowing from what appeared to be an explosion near the Pentagon went viral. Verified Twitter accounts shared it. A fake Bloomberg news account amplified it. The stock market dipped before authorities confirmed no explosion had occurred.

Pentagon in (Fake) Flames. Source: Bloomberg

Experts quickly identified it as AI-generated. The fence looked warped. The smoke seemed unnatural. And crucially, no corroborating footage existed. No one else photographed the explosion because it never happened.

Two months earlier, an image of Pope Francis in a stylish white Balenciaga puffer jacket fooled millions before being identified as a Midjourney creation. Again, the giveaway wasn't just visual inconsistencies. It was the absence of any other photograph from any other angle.

A viral AI-generated picture of the Pope wearing a puffer jacket. (Source: X).webp

These cases show when detection helps:

  • Extraordinary claims (explosions, celebrities behaving unusually, breaking news) with no corroborating evidence.
  • Single images from unverified or suspicious sources.
  • Situations where absence of other documentation is itself suspicious

Detection fails when:

  • Multiple independent sources document the same event (like the Harris rally).
  • Verification teams have already confirmed authenticity
  • The detection result contradicts abundant real-world evidence

Never use detection to dismiss well-documented events. In that context, citing AI detection tools to claim the event was fake isn't scepticism. It's denial.

How to Evaluate Suspicious Images Yourself

Photo by Geri Mis on Unsplash

Rather than relying solely on automated detectors, here's a more practical approach:

1. Reverse image search. Upload the image to Google Images or TinEye. If it's a real photograph, you'll often find the original source, the photographer's credit, or news coverage that includes the same image.

2. Check for multiple angles. Real events are photographed by multiple people. The Harris rally Trump called fake? Dozens of attendees posted their own photos and videos from different positions in the crowd. A single suspicious image with no corroborating footage deserves scepticism.

3. Look for source verification. Major news organisations like the BBC, Reuters, and AFP have verification teams. If they've confirmed an image is authentic, that carries more weight than an AI detector's probability score.

4. Examine the context. Who shared this image first? What's their motivation? An image from a verified journalist at a news outlet has different credibility than an anonymous post on social media.

5. Check obvious AI tells. While AI-generated images are improving rapidly, many still have issues with hands (wrong number of fingers, awkward positioning), text (garbled or nonsensical words), and background details (objects that melt into each other, inconsistent shadows).

These methods aren't foolproof. But they're more reliable than trusting a tool that's wrong 42% of the time.

The Solution: Voluntary Transparency

AI labeling on Facebook. Source: Meta

Rather than creating a surveillance state where every image is suspect until proven innocent, we need a cultural shift towards voluntary transparency.

Imagine if platforms made it easy for creators to flag their own AI-generated work. Not as a scarlet letter, but as a simple informational label. "This image was created with AI assistance." "This photo was enhanced using AI tools." "This is a completely AI-generated image."

Some platforms are already moving in this direction. Meta has begun labelling AI-generated images on its platforms, and many AI tools now embed invisible watermarks. But this only works if it becomes normalised rather than stigmatised.

The key word here is "voluntary". When people feel they'll be punished for using AI tools, they hide their usage. When we create an environment where AI assistance is accepted and expected to be disclosed, transparency becomes the norm.

AI labelling on Instagram. Source: Meta

Where We Go From Here

The parallel with AI writing detectors isn't coincidental. Both technologies emerged from the same impulse: the desire for certainty in an increasingly uncertain digital landscape. We want to know what's real, what's human, what's authentic.

But certainty is precisely what these detectors can't provide. The detector's model is only as good as the database it was trained on.

As AI generators evolve, detection models must constantly update to keep pace. There will always be a lag, always be errors, always be people caught in the crossfire.

We need to accept that we're entering an era where perfect detection isn't possible. What matters isn't whether an image was technically created by AI, but whether it's being used honestly.

A clearly labelled AI-generated illustration for a blog post isn't misinformation. An AI-generated photograph passed off as documentary evidence is.

Context and intent matter more than the technical means of creation.

A Final Thought

I'm not against AI image detectors existing. They have their place in the professional toolboxes of fact-checkers, journalists, and platform moderators. But the wholesale deployment of these tools as gatekeepers, with the power to brand people as frauds based on probabilistic assessments, will create more harm than good.

We watched it happen with writing. Students falsely accused. Careers damaged. Trust eroded.

Let's not make the same mistake with images.

The technology isn't the problem. Our blind faith in it is.