AI Blog
AI and Copyright: Navigating Intellectual Property

AI and Copyright: Navigating Intellectual Property

Published: April 24, 2026

AIcopyrightintellectual-property

Introduction

Artificial intelligence is reshaping how content is created, distributed, and consumed at a breathtaking pace. Generative AI tools like ChatGPT, Midjourney, and GitHub Copilot can produce text, images, music, and code in seconds — outputs that would have taken human professionals hours or even days. But beneath this technological revolution lies a legal minefield that companies, creators, and policymakers are only beginning to map.

The central question is deceptively simple: Who owns content created by AI? And the follow-up is equally thorny: Is it legal for AI to learn from copyrighted material in the first place?

These are not hypothetical questions. Lawsuits are already being filed, legislation is being drafted, and billion-dollar businesses hang in the balance. In 2023 alone, more than 40 major copyright-related lawsuits involving AI were filed in the United States, and that number has continued to climb. Whether you're a developer building an AI product, a creative professional worried about your livelihood, or a business looking to leverage generative AI, understanding the intersection of AI and copyright law is no longer optional — it's essential.

This guide breaks down the key issues, real-world cases, and practical frameworks you need to navigate intellectual property in the age of AI.


What Is Copyright — And Why Does AI Complicate It?

Copyright is a form of legal protection granted automatically to the creators of original works — including text, images, music, software code, and more. Under U.S. law (and most international frameworks), copyright attaches the moment a work is "fixed in a tangible medium of expression." The creator holds exclusive rights to reproduce, distribute, display, and create derivative works from that content.

The problem with AI is that it challenges every single assumption baked into traditional copyright law:

  • Who is the author? Copyright law assumes a human creator. AI systems are not legal persons.
  • What counts as original? If an AI generates an image by statistically recombining patterns from millions of human-made images, is the output truly original?
  • Is training on copyrighted data fair use? AI models are trained on vast datasets scraped from the internet, much of which is protected by copyright.

There are two distinct legal battlegrounds here: training data copyright and output copyright, and they require very different analyses.


The Training Data Problem: Did AI "Steal" From Artists?

How AI Models Are Trained

Large language models (LLMs) like GPT-4 and image generators like Stable Diffusion are trained on datasets containing hundreds of billions of tokens — words, sentences, images, and code — much of it scraped from publicly accessible websites. For reference, GPT-4 was reportedly trained on roughly 1 trillion tokens, while LAION-5B, the dataset used to train Stable Diffusion, contains 5.85 billion image-text pairs, the majority sourced from the public internet.

The core legal issue: most of those images, articles, books, and code snippets are protected by copyright. The AI companies argue this is fair use — a legal doctrine that permits limited use of copyrighted material without permission under certain conditions. Rights holders argue it's straightforward infringement.

Real-World Example: Getty Images vs. Stability AI

In January 2023, Getty Images filed a landmark lawsuit against Stability AI (the company behind Stable Diffusion) in both the U.S. and UK. Getty alleged that Stability AI illegally scraped more than 12 million photographs from its platform — along with their associated metadata and watermarks — to train its image generation model.

Perhaps the most damning evidence? Users could prompt Stable Diffusion to produce images that included distorted versions of the Getty Images watermark, a visual artifact suggesting the model had memorized parts of the training data. This type of behavior — called "memorization" or "training data extraction" — is a significant legal and technical concern.

The case is still ongoing, but it has already sent shockwaves through the industry, prompting many AI companies to audit their training pipelines.

Real-World Example: The New York Times vs. OpenAI

In December 2023, The New York Times filed a blockbuster lawsuit against OpenAI and Microsoft, alleging that millions of its articles were used without permission to train GPT models. The Times presented evidence showing that ChatGPT could reproduce long verbatim passages from its paywalled articles — a behavior that undermines the Times's subscription business model.

This case is particularly significant because the Times is not a small creator — it's a powerful media institution with deep legal resources, and its lawsuit directly challenges the "fair use for AI training" argument that OpenAI and Microsoft have relied on.


The Output Problem: Who Owns What AI Creates?

Even if we set aside the training data question, a second thorny issue remains: who owns the copyright on AI-generated outputs?

The U.S. Copyright Office's Position

The U.S. Copyright Office (USCO) has been remarkably consistent on this: copyright requires human authorship. In a series of rulings since 2022, the USCO has rejected copyright applications for AI-generated content.

Most famously, in the case of Kristina Kashtanova and the graphic novel Zarya of the Dawn, the Copyright Office initially registered the work, then partially rescinded it. The text — written by a human — remained protected. But the images, generated by Midjourney, were declared unprotectable. The ruling established a key principle: AI-generated elements of a work cannot be copyrighted, but human-authored elements within the same work can be.

This creates a patchwork of protection for hybrid human-AI creative works — a rapidly growing category.

What This Means for Businesses

If you're a company using AI to generate marketing copy, product images, or software code, you may have no copyright protection over that output. A competitor could theoretically copy your AI-generated assets without legal consequence. For businesses investing heavily in AI-generated content, this is a significant strategic risk.


International Perspectives: The World Is Not Aligned

Different countries are taking wildly different approaches to AI copyright, creating a complex global landscape.

Country Training Data AI-Generated Output Key Development
United States Fair use debate ongoing Not copyrightable (USCO position) NYT v. OpenAI; USCO guidelines 2023–2024
European Union TDM exception (opt-out allowed) Human authorship required EU AI Act (2024) addresses transparency
United Kingdom Computer-generated works protected Author = "person who made arrangements" Law Review underway (2024)
Japan Very permissive for AI training Generally not protected Government pro-AI policy stance
China Unclear, evolving Some protection if human involvement shown Interim AI regulations (2023)
Canada Fair dealing debate No clear framework yet Consultation ongoing

Japan stands out as notably AI-friendly: its 2019 amendment to the Copyright Act explicitly allows AI companies to use copyrighted material for machine learning without permission, making it a favorable jurisdiction for AI development. The EU, by contrast, offers rights holders an "opt-out" mechanism under its Text and Data Mining (TDM) exception — meaning copyright holders can explicitly prohibit the use of their content for AI training.


Practical Frameworks for Businesses and Creators

For Businesses Using Generative AI

1. Audit Your AI Tools' Training Data Before deploying a generative AI tool, investigate what data it was trained on. Some providers — like Adobe Firefly — have been explicit that their models are trained only on licensed content, Adobe Stock images, and public domain material. This "clean" approach significantly reduces legal risk.

2. Use Commercially Licensed Models Several AI providers now offer indemnification clauses — legal promises to cover you if their AI's output leads to a copyright infringement claim. OpenAI's enterprise tier and Microsoft's Copilot Copyright Commitment are examples. These don't eliminate risk but shift liability.

3. Document Human Contributions To maximize copyright protection over AI-assisted works, document the human creative decisions involved — the prompts, edits, selection processes, and modifications. The more human authorship you can demonstrate, the stronger your copyright claim.

For Creators Protecting Their Work

1. Use Technical Opt-Out Tools Tools like Spawning's "Have I Been Trained?" allow artists to check whether their work appears in major training datasets like LAION and opt out of future training. Similarly, Glaze (developed by the University of Chicago) applies imperceptible perturbations to images that disrupt how AI models learn from them — a kind of digital armor for artists.

2. Watermark and Register Your Work While not foolproof, registering your work with the Copyright Office (in the U.S.) strengthens your legal standing in any infringement action. Watermarking — even digital, invisible watermarking — can help establish provenance.

3. Stay Informed Through Reliable Resources The legal landscape is evolving monthly. To build a solid foundation, consider reading intellectual property and technology law books that cover the intersection of digital innovation and legal frameworks.


The Emerging "AI Licensing" Economy

One promising development is the emergence of AI licensing markets — formal systems where rights holders can license their content specifically for AI training at negotiated rates.

  • Getty Images launched an AI training dataset product in 2023, allowing companies to license its image library for AI training purposes legally.
  • The Associated Press signed a licensing deal with OpenAI in 2023 for access to its news archive.
  • Shutterstock partnered with OpenAI to provide licensed training data, while simultaneously launching a compensation fund for contributors whose work was used.

These deals suggest a possible future equilibrium: AI companies pay for access to high-quality training data, and rights holders receive compensation. But the dollar amounts remain small relative to the scale of AI training operations, and many creators remain skeptical.

For a deeper dive into how markets and law interact around digital content, [digital copyright and creative economy books](https://www.amazon.

Related Articles