Skip to content

Claude Sued for Alleged Copyright Violations at OpenAI

Large language model accused of using illicit materials for training: Claude allegedly utilized pirated books in its development process.

Copyright Infringement Lawsuit Filed Against OpenAI and Claude for Alleged Plagiarism
Copyright Infringement Lawsuit Filed Against OpenAI and Claude for Alleged Plagiarism

In a landmark decision, AI company Anthropic has reached a preliminary settlement with authors who accused the company of mass copyright infringement. The settlement, expected to be finalized by September 3, comes after Anthropic was found to have downloaded millions of copyrighted books from piracy sites, including Library Genesis, to build a permanent "central library" of training data for its Claude large language model.

The settlement amount for Anthropic is widely believed to be far less than what proper licensing would have cost upfront. This marks a significant milestone, as it is the first time a generative AI company has been held financially accountable for building on the backs of unpaid creators.

U.S. District Judge William Alsup certified a class action against Anthropic, and the implications of the settlement are arguably more transformative than recent AI news like ChatGPT or Google's Gemini. A cascade of settlements is expected in the coming months as companies realize the cost of fighting and losing copyright infringement cases could be existential.

The Association of American Publishers stated that AI can't exist without using human authorship. Fair use is not a blanket defense when the method of data acquisition involves clear-cut piracy. Anthropic, along with other AI companies like OpenAI, Meta, and Google, have claimed their use of copyrighted works is transformative and therefore protected under fair use law. However, the Anthropic settlement suggests otherwise.

Legal clarity is emerging, and the message is clear: if AI is truly transformative, its foundations must be built on respect, not theft. AI ethics and compliance are moving from optional to existential for generative AI firms. The settlement between Anthropic and authors sets a precedent for ongoing lawsuits against these companies.

The systematic downloading of copyrighted books by Anthropic undermines the industry's credibility and legal position. Statutory damages for each infringing work could start at $750, and for willful infringement, could reach up to $150,000 per work. Had the case gone to trial, Anthropic faced potential damages exceeding $1 trillion.

This settlement may mark the beginning of the end of the Wild West era of AI data scraping. Companies must now consider the ethical and legal implications of their data acquisition strategies, or face the potential consequences. The Anthropic case serves as a stark reminder that respect for intellectual property is not just a matter of morality, but also a matter of survival for AI companies.

Read also:

Latest