Anthropic’s AI copyright settlement sets a legal marker on fair use vs piracy, reshaping training data practices and accelerating licensing and regulation.

Anthropic has reached the first major settlement in the wave of copyright lawsuits targeting AI companies, potentially setting a precedent that could fundamentally alter how the industry sources and handles training data. The San Francisco-based company settled a class action lawsuit brought by authors alleging unauthorised use of copyrighted books to train its Claude AI model.
The case, Bartz v. Anthropic, was filed in August 2024 by authors Andrea Bartz, Charles Graeber and Kirk Wallace Johnson. What distinguished this lawsuit from others was its focus on how Anthropic obtained the training materials rather than just their use.
In June, US District Judge William Alsup delivered a split decision that illuminated the complex legal terrain surrounding AI training data. The judge ruled that Anthropic’s use of copyrighted books for AI training constituted fair use, providing some relief to the industry. However, Alsup found that the company violated copyright law by downloading approximately seven million books from pirate sites like LibGen.
This distinction between legitimate AI training and data piracy represents a crucial legal boundary. While companies can potentially claim fair use for training purposes, they cannot rely on illegally obtained source materials. The ruling suggests that AI developers must be more careful about data provenance, especially when working with large datasets that may include pirated content.
The potential financial exposure was staggering. With damages potentially reaching billions or even exceeding $1 trillion, Anthropic faced catastrophic liability ahead of the scheduled December trial. The settlement, reached this month, prevents what could have been a precedent-setting damages award.
AI companies face mounting legal challenges over training data practices. Consolidated lawsuits in New York involve OpenAI, Microsoft and Meta, with prominent authors like John Grisham and Jonathan Franzen, alongside major news publishers including The New York Times, seeking damages and injunctions.
Subscribe to our newsletter and never miss a story. No spam, ever.

The president accepted a 10-point peace plan that gives Iran nearly everything it asked for. Hours later, he contradicted its central demand. Either he did not read it or he does not care what it says.

Anthropic seals off the last third-party route into its Claude subscription tier, forcing OpenClaw and all other AI agent platforms onto metered billing.

A debugging file left in a software update exposed 512,000 lines of source code, 44 unreleased features, and a mode that hides AI involvement in open-source projects. It was Anthropic's second data exposure in a week.
The resolution points toward more structured agreements between tech companies and content creators. As visualisation tools like Lumion software demonstrate in architecture and design, the ability to process and render complex datasets requires careful attention to source materials and licensing – a principle now extending to AI training datasets.
Edward Lee, a law professor at Santa Clara University who follows AI copyright litigation, noted the significance of Anthropic’s decision to settle rather than face trial. The company had recently hired a new trial team, suggesting preparation for a lengthy legal battle before opting for resolution.
While settlement terms remain undisclosed, the agreement reportedly includes provisions for author compensation, transparency requirements and opt-out mechanisms for future use of their work. These elements could become standard features in similar settlements, establishing new norms for how AI companies handle copyrighted content.
The Authors Guild is coordinating notifications to affected writers, highlighting the class action’s broad scope. Court finalisation is expected in early September, marking the conclusion of what many consider the first settlement in a string of major industry lawsuits.
The resolution may accelerate licensing deals between tech companies and content creators as an alternative to litigation. Recent months have seen increased discussion about structured licensing arrangements that could provide predictable revenue streams for authors and publishers while giving AI companies clear legal access to training materials.
Beyond individual cases, the settlement could influence broader regulatory approaches to AI training practices. Companies face increasing scrutiny over data sourcing, and the distinction between fair use training and content piracy established in this case may inform future regulations.
The Brazilian newspaper Folha de S.Paulo’s recent copyright lawsuit against OpenAI demonstrates that legal challenges are expanding globally, with international publishers and content creators pursuing similar claims. This mirrors the regulatory pressure facing other tech giants in multiple jurisdictions.
As AI companies navigate this evolving legal terrain, the Anthropic settlement represents more than just case resolution. It marks a fundamental shift toward regulated and transparent practices that must balance technological advancement with intellectual property rights. The outcome could well define how the AI industry approaches training data for years to come, particularly as privacy concerns around AI data collection continue to mount across different sectors.

London Tech Week returns to London Olympia from 8 to 12 June with a new Deep Tech Stage spanning quantum computing, space, surgical robotics and life sciences.