Photo Verification

How to Tell If a Photo Is AI-Generated: Detection Tools vs. Provenance Verification

A practical guide to determining whether a photo is real or AI-generated. Compares AI detection tools with provenance-based verification and explains why verifiable history outperforms probability scores.

ByLumethic Team
10 min read
Share

The question comes up constantly now. A striking image appears in a news feed, a stock library, or a contest submission, and someone asks: is this real? The tools for generating synthetic images have become good enough that the old visual tells (malformed hands, nonsensical text, uncanny skin texture) are disappearing. For anyone who works with photography professionally, or anyone who simply wants to know whether a picture shows something that actually happened, this is a practical problem that needs a practical answer.

There are two fundamentally different ways to approach it. One is reactive: take a finished image and try to determine, after the fact, whether it was made by a machine. The other is proactive: establish a verifiable record of where an image came from and how it was created, so that authenticity can be confirmed rather than guessed at. Both have their uses, but they work very differently, and understanding the distinction matters.

Two Fundamentally Different Approaches

AI detection tools analyze an image's pixels and try to identify statistical patterns left behind by generative models. They are classifiers: they take an image as input and return a probability score. The output is a guess, sometimes a well-informed guess, but a guess nonetheless.

Provenance verification takes the opposite approach. Rather than analyzing an image for signs of synthetic origin, it checks the image against its source material, typically the original RAW file from the camera sensor. If the image can be computationally linked back to genuine camera data, its authenticity is established through evidence, not inference.

The difference is structural. Detection asks "does this look fake?" Provenance asks "can we prove this is real?"

How AI Detection Tools Work

Most AI image detectors are trained on large datasets containing both photographs and synthetic images. The model learns to recognize patterns that distinguish one from the other: subtle artifacts in color distribution, texture consistency, frequency-domain signatures, or noise patterns that differ between camera sensors and neural networks.

When you submit an image, the detector runs it through this trained model and returns a confidence score. A typical result might read "87% likely AI-generated" or "93% likely authentic." Some tools provide additional analysis, such as heat maps highlighting regions of the image that triggered suspicion.

Several detectors are publicly available. Tools like Hive Moderation, Illuminarty, and AI or Not offer web-based analysis. Adobe's Content Authenticity initiative takes a different approach, providing inspection tools that read C2PA provenance metadata rather than classifying pixels. Research groups at universities and labs continue to publish new approaches as generative models evolve.

These tools can be genuinely useful in certain situations. When you have no other information about an image and need a quick initial assessment, a detection tool provides a starting point. For content moderation teams processing thousands of uploads, automated detection at scale serves as a first filter.

Where AI Detectors Fall Short

The limitations of detection tools become apparent quickly in professional contexts where accuracy matters.

The core problem is the arms race. Detectors learn to spot artifacts that current generative models produce. When a new model version eliminates those artifacts, the detector's accuracy drops until it is retrained. This cycle has no endpoint. Each generation of image synthesis closes the gap, and detectors must constantly catch up. Studies have shown that detectors trained on GAN-era images perform poorly on diffusion model outputs, and detectors tuned for Stable Diffusion struggle with newer architectures.

False positives are a serious and underappreciated issue. Real photographs are regularly flagged as AI-generated, particularly images with certain characteristics: studio lighting, shallow depth of field, extensive post-processing, or subjects that happen to match patterns the detector associates with synthetic content. Real photographs are increasingly being mistaken for AI output. An Australian photographer had a genuine iPhone capture rejected from a photo contest by judges who deemed it "a little AI-ish," and similar incidents are becoming more common as the visual quality of synthetic images converges with real photography.

False negatives are equally problematic in the other direction. Simple modifications to an AI-generated image, such as screenshotting it, applying a filter, or re-saving at a different compression level, can be enough to fool many detectors. The image is still synthetic, but the statistical fingerprints the detector relies on have been disturbed.

The opacity of the output is another limitation. A probability score tells you nothing about why the image was flagged. If a detector returns "78% likely AI," you have no way to evaluate that claim independently. There is no supporting evidence, no chain of reasoning, and no way for a third party to audit the conclusion. In legal, editorial, or forensic contexts, a percentage is not useful evidence.

Finally, detectors provide no information about provenance. Even if a detector correctly identifies an image as a real photograph, it cannot tell you who took it, when, with what equipment, or whether it has been tampered with since capture. It answers one narrow question (real or synthetic?) and leaves everything else unknown.

How Provenance Verification Works

Provenance verification starts from a different premise. Instead of trying to classify an image based on pixel analysis alone, it establishes a verifiable link between the finished image and its source.

For photography, the strongest form of provenance is RAW file verification. A camera's RAW file contains the unprocessed data from the sensor, including Bayer pattern information, sensor noise characteristics, and device-specific metadata. This data is extremely difficult to fabricate convincingly. When a photographer provides both their finished JPEG and the original RAW, a verification system can run a series of forensic comparisons to determine whether the JPEG is a legitimate derivative of that RAW file.

These comparisons operate across multiple independent dimensions. Sensor authenticity checks examine whether the RAW file exhibits characteristics consistent with genuine camera hardware. Structural similarity analysis measures whether the visual content of the JPEG corresponds to the RAW at a perceptual level. Histogram analysis compares the statistical distribution of color and luminance values. Metadata consistency checks look for discrepancies between the technical parameters recorded in the two files. Additional checks for recapture artifacts, face region integrity, and perceptual hash alignment provide further layers of evidence.

Because these verification methods are independent and examine different signal types, defeating them simultaneously is significantly harder than fooling a single-model classifier. The output is not a probability score but a concrete verification report, documenting which checks passed, what evidence was found, and how the two files relate to each other.

When verification succeeds, the system can sign the JPEG with a C2PA manifest, a cryptographic certificate that records the verification results, the identity of the signer, and a timestamp. This manifest travels with the image and can be inspected by anyone downstream. It transforms "we believe this photo is real" into "here is the evidence, signed and timestamped, that this photo derives from a verified camera file."

Detection vs. Provenance in Practice

Consider a concrete scenario. A news organization receives a photograph from a freelancer covering a breaking story. The editor needs to know whether the image is genuine before publication.

Using an AI detector, the editor uploads the image and receives a confidence score. If the score reads "91% authentic," that sounds reassuring, but the editor has no way to verify that number. If the score reads "65% authentic," the image might still be completely real, just with characteristics the model finds ambiguous. The editor is left to make a judgment call based on a number they cannot interrogate.

Using provenance verification, the editor asks the freelancer to submit the original RAW file alongside the JPEG. The verification system runs its multi-factor analysis and produces a detailed report. The editor can see that the RAW file passes sensor authenticity checks, that the JPEG shows high structural similarity to a normalized rendering of the RAW, that metadata is consistent between the two files, and that no recapture artifacts were detected. The image is then signed with a C2PA manifest. If questions arise later, the evidence is on record.

The two approaches differ in what they demand. Detection requires only the image itself, which is convenient. Provenance requires the source file, which is more demanding but produces stronger evidence. This tradeoff is central to choosing the right approach for a given context.

When to Use Which Approach

AI detection tools are most useful when you have no access to source material and need a quick, approximate assessment. Content moderation at scale, initial triage of user-uploaded images, and informal curiosity about a specific picture are all reasonable use cases. The key is to treat the output as a signal, not a verdict.

Provenance verification is the appropriate tool when the stakes are higher: editorial publication, legal evidence, insurance claims, contest judging, stock photography licensing, academic research imagery, or any context where "probably real" is not good enough. In these cases, the ability to produce a documented, auditable chain of evidence matters. It also matters that the verification results can be attached to the image permanently via C2PA credentials, so that downstream consumers of the image can independently verify its status.

Many workflows benefit from combining both. A quick AI detection check can flag images that warrant closer examination. Provenance verification then provides the definitive assessment for anything that matters.

How Lumethic Approaches This Problem

Lumethic is built around provenance verification. The platform compares a photographer's JPEG to the original RAW file using eight independent forensic analysis techniques: sensor authenticity verification, EXIF metadata validation, structural similarity measurement, perceptual hash comparison, histogram analysis, face detection and comparison, RAW integrity assessment, and recapture detection.

All eight checks must pass before the system will sign the image. This consensus requirement means that a single weak link does not compromise the overall result. When verification succeeds, Lumethic generates a C2PA manifest and embeds it in the JPEG, creating a permanent, inspectable record of the image's verified status.

The RAW file is used for analysis and then deleted. It is never stored on Lumethic's servers. This matters because the RAW file is the photographer's most sensitive asset, and any system that asks for it must handle it responsibly.

For photographers working in Adobe Lightroom, the Lumethic Lightroom plugin integrates verification directly into the export workflow. For mobile photographers, the Lumethic Capture iOS app creates verified images at the point of capture. For organizations that need to verify images programmatically, the Lumethic API supports automated batch processing.

The free tier includes five verifications per month, which is enough to test the workflow on your most important images. Try it here.

Frequently Asked Questions

Can an AI-generated image pass provenance verification? No, because provenance verification requires a genuine camera RAW file. AI-generated images do not originate from a camera sensor and therefore have no corresponding RAW file. Without a RAW file that passes sensor authenticity checks and matches the submitted image, verification will fail.

How accurate are AI detection tools in 2026? Accuracy varies significantly by tool, by the generative model used to create the image, and by any post-processing applied. Published benchmarks often reflect controlled lab conditions. In real-world use, with images that have been compressed, cropped, filtered, or re-saved, accuracy can be substantially lower. False positive rates (genuine photos flagged as AI) remain a persistent problem.

Do I need a special camera for provenance verification? No. Any camera that shoots RAW is compatible. You do not need a camera with built-in C2PA support. Lumethic's verification works by analyzing the RAW file you already produce as part of your normal shooting workflow.

What is C2PA? C2PA (Coalition for Content Provenance and Authenticity) is an open technical standard for embedding provenance information into digital content. A C2PA manifest is a cryptographically signed record that travels with the image, documenting its origin, verification status, and edit history. For a detailed explanation, see What is C2PA?.

What happens to the provenance data when I share an image on social media? Most social media platforms currently strip embedded metadata, including C2PA manifests, during their upload and compression process. However, this is changing. Google now surfaces C2PA data in its "About this image" feature across Google Images and Lens. As more platforms adopt C2PA support, provenance data will increasingly survive distribution.

Can I use both detection tools and provenance verification? Yes, and for many workflows this is a reasonable approach. AI detection can serve as a quick initial screen, while provenance verification provides definitive evidence for images that require it.


Related Reading

#AI Detection#Provenance#C2PA#Photo Verification#Authenticity