
.png)
If your content is on the internet, it has probably been ingested by an AI training pipeline. This is not speculation — it's the documented practice of every major AI lab. The question isn't whether your content was used. It's whether you have any mechanism to assert that it shouldn't have been.
Most organizations don't. Here's what actually works — and what doesn't.
Updating your terms of service to prohibit AI training is a common first step. It's also largely unenforceable. AI training happens at scale through automated pipelines that don't read terms of service. Even if a violation could be proven, litigation is slow, expensive, and uncertain. Terms of service are important for legal record-keeping, but they're not a technical control.
The robots.txt file can instruct crawlers not to index specific pages. Some AI companies have said they will honor these directives. Others have not. robots.txt is voluntary compliance — it works until it doesn't, and you have no way to verify whether it's being respected.
Some organizations block known IP ranges associated with AI scraping. This is a cat-and-mouse game. IP ranges change. Proxies exist. Content already scraped before the block remains in training datasets.
The C2PA standard includes a formal mechanism for asserting "Do Not Train" — a cryptographically signed assertion embedded in the content file itself that travels with the asset through distribution.
This is categorically different from robots.txt. The signal isn't a request to a crawler — it's a verifiable, tamper-evident declaration embedded in the content. When an AI system ingests content carrying a signed "Do Not Train" assertion, that's no longer an accident. It's a documented decision to override an explicit rights signal.
That distinction matters legally. It removes plausible deniability.
C2PA metadata can be stripped — screenshots don't carry metadata, and many platforms remove it on upload. Imperceptible watermarking addresses this by embedding provenance signals into the content itself, not just the file wrapper.
A watermark interwoven with an image, video, or audio file survives compression, cropping, format conversion, and reposting. Even when metadata is gone, the signal remains. Detection tools can identify the original creator regardless of how many times the content has been copied.
For text content, watermarking isn't practical — you can't invisibly mark words. The alternative is a digitally signed provenance certificate: a manifest that formally records the origin and rights status of a piece of text, issued at creation and verifiable independently of the document itself.
Effective content protection isn't a single technology — it's a stack:
This is how enterprise content protection actually works. It doesn't stop every bad actor, but it creates a verifiable record that makes unauthorized AI training legally distinguishable from authorized use.
Governments are reinforcing this framework. The EU AI Act, effective August 2026, requires C2PA-compliant provenance for AI-generated and AI-modified content. South Korea and India have enacted labeling requirements. The direction is clear: technical provenance signals are becoming mandatory, not optional.
Organizations that implement now aren't just protecting their content. They're building the compliance infrastructure they'll need anyway.
Limbo handles the entire stack — C2PA metadata generation, watermarking, provenance certificates — as API-first infrastructure that integrates with your existing workflows. Request a demo to see it in action.
.png)