Protecting your creative work from unauthorized AI training requires a multi-layered approach combining legal protections, platform settings, and technical measures. The most direct actions you can take today include opting out of AI training on platforms where you share content, using robots.txt to signal to AI crawlers that your website shouldn’t be indexed, and deploying cloaking tools like Glaze or Nightshade to confuse AI models attempting to scrape your work. Additionally, new regulations—particularly the EU AI Act (effective 2026) and emerging U.S. legislation—are creating legal frameworks that require AI companies to respect copyright opt-outs and disclose their training data sources.
The landscape shifted dramatically in 2025-2026 as major copyright settlements and lawsuits raised the financial stakes for AI companies using creator content without permission. Anthropic agreed to pay $1.5 billion to authors in one of the largest copyright settlements in U.S. history, while Universal Music Group and Warner Music Group secured licensing deals protecting musical works. These settlements indicate that courts and regulators are increasingly supporting creators’ rights. This article covers the regulatory protections now in place, platform-specific opt-out options, technical safeguards you can implement, what recent settlements mean for your rights, and how new compensation models may reshape creator protection in the years ahead.
Table of Contents
- What New Laws Require AI Companies to Disclose and Respect Your Copyright
- Opting Out on OpenAI, Claude, LinkedIn, and Google Platforms
- Using Robots.txt and Blocking AI Crawlers from Scraping Your Work
- What the Anthropic Settlement and Music Industry Deals Tell You About Your Rights
- The Gap Between Legal Protections and Actual Security
- A Proposed Alternative: The “Learnright” Model for Creator Compensation
- The Investment Angle: How Creator Rights Are Reshaping AI Industry Dynamics
- Conclusion
- Frequently Asked Questions
What New Laws Require AI Companies to Disclose and Respect Your Copyright
starting in 2026, the EU AI Act is fundamentally changing how AI developers can source training data. Under this regulation, every AI company must disclose where they sourced their training data, respect copyright opt-outs when creators have explicitly reserved their rights, and label any AI-generated content they produce. This means AI developers are now required to check whether a data source has a copyright reservation before using it—a verification step that didn’t exist in earlier AI development. For creators in European jurisdictions, this creates a binding legal requirement that AI companies cannot ignore without facing regulatory penalties.
Beyond the EU, the landscape is shifting in the United States as well. The bipartisan TRAIN Act, introduced in early 2026, seeks to empower creators with legal transparency and meaningful access to courts when their copyrighted work is used for AI training without authorization. While still proposed, this legislation reflects growing political consensus that creators deserve protection. The EU Parliament has gone even further, urging the creation of a European register at the European Union Intellectual Property Office (EUIPO) that would list every copyrighted work used to train AI models and document which artists have opted out—essentially creating a public database that AI companies must check before training their systems.

Opting Out on OpenAI, Claude, LinkedIn, and Google Platforms
If you create content on ChatGPT, LinkedIn, or other AI platforms, you have immediate control over whether your specific conversations or posts are used to train AI models. OpenAI users can disable model training by navigating to Settings > Data Controls and toggling off “Improve the model for everyone,” or by using OpenAI’s privacy portal to submit a “do not train on my content” request. claude (Anthropic) allows users as of September 2025 to toggle whether their conversations are used for AI training directly in settings. Google Gemini users can disable training through Activity settings and delete conversation data to prevent future model improvements from incorporating their input.
LinkedIn took a significant step beginning November 3, 2025, by offering users the option to opt out of having their data used to train content-generating AI models. However, a critical limitation exists here: opting out of training on these platforms only protects content you create within those platforms. Your independently published blog posts, articles, artwork, or music hosted on your own website or other platforms are not covered by these toggles. You must take additional steps—such as using robots.txt or technical cloaking tools—to protect content outside these major platforms. The opt-out feature protects you where you create, but creators publishing across multiple channels need a broader protection strategy.
Using Robots.txt and Blocking AI Crawlers from Scraping Your Work
The simplest technical protection available to any content creator with a website is robots.txt, a standard file that signals to AI crawlers—including OpenAI’s GPTBot, OAI-SearchBot, Anthropic’s ClaudeBot, and Perplexity’s PerplexityBot—that your content should not be indexed or used in AI training. Adding clear directives to your robots.txt file can tell these bots to stay off your site entirely. However, there is a critical limitation you must understand: robots.txt is not legally binding. It operates on a voluntary-compliance basis, relying on reputable AI companies to respect the signal. While major AI companies generally honor robots.txt for reputational reasons, deliberately ignoring it may strengthen your legal case if you later need to claim bad faith or trespass, but it won’t automatically prevent a bad actor from scraping your content.
More advanced creators are turning to cloaking tools like Glaze and Nightshade, developed at the University of Chicago. Glaze makes tiny, imperceptible pixel-level changes to your images that confuse AI models attempting to learn your artistic style, making it harder for the model to reproduce your unique creative voice. Nightshade takes this a step further by making artwork “poisonous” to AI training—the tool embeds subtle, invisible changes into images that don’t affect human perception but degrade the quality of AI models trained on poisoned images. These tools are still emerging technologies, and their effectiveness depends on the specific AI training method and whether the model was trained after the poisoning was applied. They offer protection for visual creators but are less applicable if you work primarily in text, music, or other formats.

What the Anthropic Settlement and Music Industry Deals Tell You About Your Rights
The $1.5 billion settlement Anthropic agreed to pay authors in September 2025 represents a watershed moment for creator rights. This was one of the largest copyright settlements in U.S. history, and the figure is striking because Anthropic’s potential exposure was theoretically in the hundreds of billions of dollars. The settlement signals to courts, regulators, and other AI companies that using copyrighted training data without explicit permission carries severe financial consequences. For creators, this demonstrates that copyright courts are increasingly willing to side with artists and authors, even against well-funded AI companies backed by billions in venture capital.
The music industry secured similar wins. Universal Music Group and Warner Music Group both negotiated settlements in 2025 that include plans to launch new subscription services in 2026 for generative AI trained on fully authorized and licensed music. Rather than allowing AI companies to scrape music freely, these labels forced AI developers to negotiate licensing deals where artists can opt in and receive compensation. Meanwhile, the legal threats continue: in January 2026, Universal Music Publishing Group, Concord Music, and ABKCO filed a $3.1 billion lawsuit against Anthropic alleging unauthorized use of torrented piracy. These suits and settlements create a legal precedent that directly protects musicians and other creators—if an AI company uses your work for training without a license, you may have grounds for substantial damages.
The Gap Between Legal Protections and Actual Security
While regulations and opt-outs provide important safeguards, they are not foolproof. The most significant gap is enforcement: the EU AI Act requires disclosure of training data sources and respect for opt-outs, but ensuring that an AI company actually complied requires investigation, legal action, or regulatory audits. A company could theoretically claim compliance while secretly scraping content. Similarly, robots.txt and opt-out settings work only against honest actors; they provide no protection against rogue developers or companies operating in jurisdictions that ignore EU and U.S. regulations.
If a company is willing to face lawsuits or regulatory fines, technical and platform-based protections may not stop them. Another limitation: many of these protections are reactive rather than preventive. The Anthropic settlement happened after millions of words were already used to train Claude. Opting out now does not un-train the models that were already built. The New York Times’ lawsuit against Perplexity for copyright infringement over illegal crawling illustrates this point—even with all available protections, determined AI companies may still attempt to scrape content, and you must then pursue legal action to seek damages or removal. For creators whose work has high economic value or who fear their style will be copied by AI-generated content, relying solely on these defenses may feel insufficient, which is why many are turning to technical cloaking tools like Glaze and Nightshade as insurance, even though their long-term effectiveness remains uncertain.

A Proposed Alternative: The “Learnright” Model for Creator Compensation
Rather than relying entirely on prohibition and opt-outs, a new legal framework is emerging: the “Learnright” model. This approach proposes giving creators the explicit right to license their work specifically for AI training with direct compensation, similar to how musicians license songs for radio play or streaming. Under this model, creators would no longer face the legal uncertainty of whether AI companies can use their work without permission. Instead, they would have a clear, licensed pathway to allow AI training in exchange for payment.
This removes ambiguity and creates a potential revenue stream from the same work that might otherwise be scraped for free. The Learnright approach addresses a fundamental issue with current protections: they are designed to prevent or penalize use, but they don’t create a mechanism for fair compensation when creators might actually want to participate. For a professional writer, musician, or visual artist, the choice between “forbid AI training entirely” and “let it happen for free” is binary and unsatisfying. A licensing model would expand that choice to include “allow AI training for a fee I negotiate.” This framework is still emerging, but it represents a future where creator protections shift from defensive blocking toward active monetization and control.
The Investment Angle: How Creator Rights Are Reshaping AI Industry Dynamics
From an investing perspective, the explosion of creator rights lawsuits and regulations is reshaping the AI landscape in 2026. AI companies that anticipated and planned for licensing costs and settlements (like those now paying royalties for music) will have more sustainable business models than those betting on unrestricted, free access to training data. The billion-dollar settlements demonstrate that content licensing is now a material cost for AI developers, not an optional expense. Companies building AI systems without a clear plan for addressing copyright claims face enormous tail risk. For individual creators and small creative businesses, the current moment offers clarity on what to do now and hope for the future.
Implementing robots.txt, using platform opt-out settings, and considering technical tools like Glaze create immediate protection. Newer EU regulations provide legal teeth to those protections. And the Learnright model and emerging licensing deals suggest a future where creators are compensated rather than simply protected from use. The landscape is shifting in your favor, even though the protections are incomplete. As an investor or creator navigating this space, understanding these layered protections—legal, platform-based, and technical—is essential to managing both the risks and opportunities in AI’s relationship with creative work.
Conclusion
Protecting your creative work from unauthorized AI training involves immediate actions you can take today and reliance on emerging regulatory frameworks that are still developing. Start by opting out of AI training on platforms where you share content, implement robots.txt on your website, and consider cloaking tools like Glaze or Nightshade if your work is primarily visual. Understand that these protections are not foolproof—they rely on voluntary compliance or legal enforcement—but they create meaningful barriers and establish your intent to protect your work. The legal landscape is shifting rapidly in your favor.
The EU AI Act, proposed U.S. legislation like the TRAIN Act, and the explosion of copyright settlements demonstrate that regulators and courts are increasingly supporting creator rights. The $1.5 billion Anthropic settlement and music industry licensing deals show that AI companies now face severe financial consequences for unauthorized use. Look ahead to future models like “Learnright,” which may eventually allow you to license your work for AI training and receive compensation. In the meantime, take the steps available today, monitor regulatory developments in your jurisdiction, and prepare for a future where creator compensation and control become the norm rather than the exception.
Frequently Asked Questions
If I add robots.txt to my website, will AI companies definitely not use my content?
No. Robots.txt is a voluntary standard that reputable AI companies honor, but it is not legally binding. Major companies like OpenAI and Anthropic generally respect robots.txt, but a determined bad actor could ignore it. However, deliberately ignoring a clear robots.txt directive may strengthen your legal case if you later pursue a copyright claim.
Will opting out of AI training on ChatGPT protect my personal blog?
No. Opting out on ChatGPT, Claude, or LinkedIn only protects content you create within those platforms. Your independently hosted blog, website, or content on other platforms requires separate protection through robots.txt, technical tools, or legal action.
Do Glaze and Nightshade work on all AI models?
Glaze and Nightshade are still emerging technologies. Their effectiveness depends on the specific AI training method and whether the model was trained after your content was cloaked. They are most effective for visual creators but offer less protection for text, music, or other formats.
What should I do if I discover my content was used to train an AI model without permission?
Document the use, check if it violates your jurisdiction’s copyright laws (particularly if you’re in the EU under the AI Act), and consult a lawyer specializing in copyright or intellectual property. The Anthropic settlement and music industry cases demonstrate that you may have grounds for substantial damages.
Is the “Learnright” model available now?
No, Learnright is a proposed future framework, not currently available. However, music licensing deals in 2025-2026 represent an early version of this concept. Over time, expect more platforms and AI companies to offer creator licensing options similar to how music streaming services compensate artists.
Do I need to do all of these protection methods at once?
No. Start with platform opt-outs (if you use OpenAI, Claude, LinkedIn, etc.) and robots.txt on your website. If your work is visual, consider Glaze or Nightshade. For high-value creative work, consult a lawyer about your specific copyright situation and jurisdiction. Most creators can achieve meaningful protection with just robots.txt and platform opt-outs.