Skip to content
Go back

Anthropic’s Use of Scanned and Pirated Books Ignites AI Copyright Controversy

Samir Badaila
Published:  at  06:55 PM
4 min read
Anthropic’s Use of Scanned and Pirated Books Ignites AI Copyright Controversy

Anthropic, the AI firm behind Claude, has come under scrutiny after a judge disclosed that it trained its models using millions of scanned books—both legally purchased and pirated—sparking renewed debates over copyright in AI development. The revelation stems from a recent court ruling that highlights Anthropic’s unconventional data acquisition: the company bought millions of used print books, destructively scanned them into digital form, and discarded the originals, while also downloading over 7 million pirated titles—5 million from Library Genesis in 2021 and 2 million from Pirate Library Mirror in 2022. Judge William Alsup ruled that digitizing purchased books for training constituted fair use under U.S. copyright law, likening it to a writer learning from reading, but deemed the use of pirated books illegal, setting the stage for a December trial to determine damages. As one of the first judicial takes on AI training and copyright, this decision fuels a growing wave of lawsuits against AI companies, but its implications remain contentious—let’s break it down.

The Scanning and Piracy Details

Anthropic’s process involved physically acquiring used books, stripping their bindings, scanning the pages, and feeding the digital copies into its Claude models, a method that destroyed the originals to avoid duplication concerns. Simultaneously, the company tapped into pirated sources, amassing a vast library despite internal reservations about legality, as noted in court documents. Judge Alsup approved the purchased-book approach as transformative, arguing it didn’t replicate or compete with the originals but created “something different.” However, he sharply criticized the pirated downloads, calling them unnecessary when lawful purchases were an option, and flagged the retention of this illicit library as a copyright violation.

The establishment might spin this as a creative workaround, but the narrative glosses over ethical gray areas. The destructive scanning—while legally clever—raises questions about cultural preservation, and the pirated haul suggests a cost-saving shortcut that undermines fair compensation for authors. The lack of transparency on how much pirated data actually trained Claude, versus being stored, leaves a gap in understanding the ruling’s full scope.

This ruling marks a pivotal moment, being among the first to dissect fair use in AI training, with Alsup’s decision potentially influencing dozens of pending cases against firms like OpenAI and Meta. The fair use win for purchased books bolsters AI companies’ defenses, framing training as a non-infringing, transformative act. Yet, the piracy rebuke opens a liability door, with potential damages up to $150,000 per infringed work, which could tally into billions given the 7 million pirated titles. Anthropic’s claim that its approach fosters creativity aligns with tech’s innovation narrative, but the judge’s skepticism about piracy necessity challenges that justification.

Skepticism is warranted. The establishment’s focus on the fair use victory might overshadow the piracy trial’s uncertainty—did the pirated books meaningfully shape Claude, or were they just a convenient archive? The ruling’s reliance on Anthropic’s intent (transformative use) rather than output (e.g., whether Claude replicates styles) sidesteps a key creator concern: the risk of market displacement. Posts found on X show mixed sentiment, with some hailing the fair use precedent and others decrying the piracy as theft, reflecting the polarized debate but offering no conclusive insight.

Implications and Caution

This could legitimize AI training on legally sourced works, easing pressure on companies to license data, but the piracy ruling signals a limit—future firms might face steeper scrutiny over data origins. The establishment might tout this as a green light for AI growth, but it risks escalating tensions with authors and publishers, who see their livelihoods threatened by uncompensated use. The December trial could set a financial precedent, deterring piracy but not resolving whether training itself requires consent.

Approach with caution. The fair use win is promising for AI innovation, but the piracy verdict looms large—monitor the trial’s outcome for clarity on damages and liability. Creators should track this case’s ripple effects, while AI developers might rethink data sourcing to avoid legal pitfalls. The ruling is a step, not a settlement, in a battle that’s far from over.

Comments

Loading comments...

Comments are powered by Facebook. By using this feature, you agree to Facebook's Cookie Policy and Privacy Policy.



Previous Post
BitChat Unveiled: Jack Dorsey’s Internet-Free Messenger Takes Flight
Next Post
Microsoft’s AI Medical Team Outperforms Human Doctors by 4x