Instagram flagged my image as ‘Made With AI.’ It’s not. Recently, Instagram has been rolling out its “Made with AI” tag which ostensibly flags AI-generated images on the platform but it has been enraging photographers after flagging photos that are not AI-generated. Instagram has revealed very little about how it detects AI content saying only it uses “industry-standard indicators.” With this in mind, PetaPixel tried to find all the specific ways the “Made with AI” sticker will be attached to an Instagram past. Which Photoshop Tools Will Trigger the ‘Made With AI’ Label on Instagram? Generative Fill Perhaps one of the biggest bug-bearers for photographers is that making a minor adjustment to an image using an AI-powered tool on Photoshop results in the post being given a “Made with AI” label. My 35mm analog photo with “Made with AI” slapped on it because I removed a tiny imperfection on the upper right-hand side. The original photo with the speck highlighted. Just like in my previous tests, when removing a speck of dust on Photoshop using Adobe’s Generative Fill tool, Instagram gives the photo a “Made with AI” tag once it is uploaded. This happens despite the fact I could have achieved the exact same result with the Spot Healing Brush Tools or Content-Aware Fill which doesn’t trigger the AI label.
Of course, you can also make big alterations to a photo using Generative Fill just like below where I used the Reference Image feature which allows you to upload a reference photo to guide Generative Fill. That also triggers the marker. I swapped my dog out for a kitten using the Reference Image feature. Generative Expand Generative Expand, a tool that appears as an option when cropping out from an image, is powered by the same AI technology as Generative Fill so it is perhaps no surprise that using this tool will also trigger a “Made with Label.” The original photo is highlighted in red. The rest was expanded using the crop tool and then filled in using the AI-powered Generative Expand tool. Notable Exceptions in Adobe Lightroom and Photoshop Generative Remove Tool There is a new Generative Remove tool in Adobe Lightroom and Adobe Camera Raw (ACR), and despite it using generative AI technology, it does not trip Instagram’s “Made with AI” label. Using the Generative AI-powered Remove tool in ACR to clean up the leaves in front of the cat. The photo, which was edited using generative AI, does not receive the AI label when uploaded it to Instagram.
A lamppost was removed from this photo using Lightroom’s Generative Remove tool but it did not trigger the AI label. Generative Remove is one of the newest and most impressive AI-powered features in Adobe Lightroom and Photoshop (via Adobe Camera Raw). Currently in early access (beta), Generative Remove uses Adobe Firefly technology to remove a brushed object from an image and replace it with new pixels that match the rest of the image. While the results aren’t always much different than something like the Spot Healing tool, Generative Remove is much more reliable and sophisticated based on initial testing. Despite using generative AI technology, an image edited with Generative Remove, as much a Firefly tool as any other, doesn’t get flagged as “Made with AI” in Instagram. Why? PetaPixel contacted Adobe for comment but digging through image metadata uncovers a likely explanation. Unlike an image edited using Generative Fill, a photo edited with Generative Remove does not get tagged with any C2PA information. C2PA data is vital to Adobe’s Content Authenticity Initiative — a growing group that Meta is notably absent from — because it enables software tools to know when an image has been edited. An image edited using Generative Fill includes C2PA information. While there is a “Firefly” tag in a file edited in Lightroom using Generative Remove inside the “Retouch Area Fill method” field, a photo with Generative Fill applied has a lot more code in its metadata. In the JUMD Type and JUMD Label fields, there are C2PA flags, plus an explicit mention of “Adobe Firefly.” There are also C2PA claim generator flags and a C2PA signature in the metadata.
On the other hand, an image edited using Generative Remove, another AI-powered tool, doesn’t. Regardless of why Generative Remove doesn’t instigate C2PA flagging, at least not yet, the inconsistencies underlie part of the issue with Meta’s “Made with AI” label. It is utterly slapdash in its execution and does nothing substantive to reduce the potential harm caused by AI-generated and edited photos. Firefly or not, Generative Remove doesn’t currently trigger any action by Instagram. Other Tools in Photoshop In PetaPixel’s tests, the following tools also did not trigger Instagram’s AI label: Neural Filters, Sky Replacement, AI-powered noise reduction, and Super Resolution. It is worth noting that these features use Adobe Sensei AI technology, not Firefly. Then again, Meta’s tag isn’t “Made with Firefly,” it’s “Made with AI.” Do AI-Generated Images Trigger the ‘Made With AI’ Label on Instagram? DALL-E OpenAI’s DALL-E image generator is a popular choice, and it will trigger Meta’s “Made with AI” label. AI images generated on DALL-E will trigger the ‘Made with AI’ label. Adobe Firefly Given how some Firefly-powered features result in Meta’s “Made with AI” tag appearing on Instagram, it should come as little surprise that an image outright generated using Firefly comes with an AI content warning.
Unsurprisingly, AI images generated on Adobe Firefly will trigger Instagram’s AI flag. Stable Diffusion That said, not all AI image generators cause the “Made with AI” tag to appear. Users can create images in Stable Diffusion, export them from the platform, and upload them to Instagram without any concerns about an AI label. Remarkably, Stable Diffusion (the online version at least) doesn’t trigger Instagram’s AI label. Meta’s AI Image Generator Instagram’s parent company Meta has its own AI image generator which we cheekily used to see if Meta will flag its own AI images. Downloading the image from Meta’s AI image generator website and uploading it to Instagram will trigger the warning. And while it does flag, a simple screenshot can bypass the label — despite Meta’s AI watermark still being visible on the image itself. A screenshot version bypasses the label despite Meta’s AI sticker visible on the image.
What Does Instagram’s ‘Made with AI’ Label Actually Achieve? Considering how easy it is to circumvent Meta’s AI-detection system — if you know how to edit photos, you can bypass Meta’s current checks — it begs the question, “What’s the point?” So far, Instagram’s “Made with AI” label has managed to incorrectly label some real, very-much-not-AI photos, sullying the reputation of respected photographers, and completely miss photos that were not only truly edited with AI, but not even done so in a way designed to camouflage the use of generative AI at all. While the need for some sort of content authenticity is real, and growing ever more essential by the day, Meta’s approach is hamfisted. All indications point to a simple metadata scrape, which is not only easy to deceive, but unreliable depending on the software and individual tools a person uses to create an image. Arguably, a C2PA-based approach makes a ton of sense, but Meta has it all backward. Instead of looking for evidence that an image has been edited using generative AI, perhaps Instagram should look for evidence that an image hasn’t been edited, and those images can get a label that verifies them as authentic. This type of technology already exists, and the Content Authenticity Initiative is hard at work developing ways to make it more accessible. If only Meta wanted to play ball. Instead, photographers have to worry about whether their legitimate images will be mislabeled. There’s no doubt that in the age of AI, it is hard to trust what we see online. Even totally legitimate images have come under undue scrutiny, long before “Made with AI” labels started appearing on Instagram. It’s a big problem that requires a thoughtful solution, and the current iteration of “Made with AI” labels on Instagram is definitely not it. A Constantly Moving Target We have endeavored to find every photo editing tool that will trigger a “Made with AI” label but there may be more and we will update this article if and when we find any. There is also significant likelihood that Meta’s seemingly rudimentary and basic AI-detection technology will undergo continual changes, meaning that some features that don’t currently trigger a “Made with AI” label may in the future. Additional reporting by Jeremy Gray.
We will be happy to hear your thoughts