by Catarina Conran
Written on 27 April 2025
Large‑scale generative AI has become possible in recent years only because billions of works—books, news articles, photographs, code, cultural heritage materials, and more—are freely available online. Yet the copyright rules that govern these works were never written with machine‑learning in mind. As Large Language Models (LLMs) seep into education, research, and everyday creativity, we urgently need clear answers to three questions:
We don’t have definitive answers to these questions just yet, as courts, policymakers, and scholars are still working to determine how existing copyright principles should govern generative AI.
Against that backdrop, this blog post unpacks the current state of copyright law in this age of AI, reviews the headline lawsuits that may help to answer some of these questions, and proposes ways to channel AI’s power toward openness and reciprocity.
Human authorship remains the bedrock of most national copyright regimes. In March of this year, the US Court of Appeals for the D.C. Circuit confirmed that a work “authored exclusively by artificial intelligence” cannot receive copyright protection (Thaler v. Perlmutter). Rather, copyright vests only in the human choices behind the final expression. Thus, while AI-assisted works can be copyrighted, copyright protection can only be extended to the portions of the work over which a human exercised creative control. Accordingly, the US Copyright Office now asks applicants to explain how they used AI and to disclaim the non‑human portions.
In Canada, the EU, and Australia, the requirement for copyright protection is framed as “intellectual creation”—again implying a human mind. Meanwhile, the UK’s Copyright, Designs and Patents Act assigns copyright to the person “who made the necessary arrangements,” although that rule is now under review.
Copyright owners have the right to prevent the unauthorized reproduction of their works. Training an LLM necessarily involves making temporary (and often permanent) copies of the training data, which infringes upon the reproduction right. Thus, AI developers rely on exceptions to copyright protection to train their LLMs.
In many jurisdictions, creators enjoy moral rights—inalienable rights to be credited for their work and to object to revision, alteration, or distortion of the work. Outside the US these are granted by the Berne Convention; in the US they exist only in narrow form under the 1990 Visual Artists Rights Act and do not cover books, music, or journalism. As AI scales, safeguarding attribution and reputational integrity—even where not legally required—remains vital to a healthy commons.