Bartz v Anthropic Settlement Gets Preliminary Approval – Key Takeaways
On Thursday, Judge Alsup of the Northern District of California granted preliminary approval of a settlement in the class action […]
On Thursday, Judge Alsup of the Northern District of California granted preliminary approval of a settlement in the class action […]
On Friday, both sides in the Bartz v. Anthropic lawsuit filed motions for the court to consider regarding preliminary approval of a settlement in the class action copyright infringement lawsuit filed against Anthropic last year by three book authors.
As readers are likely aware, the Bartz v. Anthropic AI lawsuit had a couple of major developments recently. Though the lawsuit was initially brought to address the legality of using copyrighted materials for training AI, the suit has now shifted its focus to Anthropic’s storage—without training use—of copies of books downloaded from LibGen and PiLiMi, two sites that share pirated copies of books and other materials.
Authors are navigating change when it comes to copyright and artificial intelligence. We’re committed to developing and sharing practical resources
Last month, a diverse set of stakeholders gathered at New York University Law School to contribute to an implementation plan
Yesterday, Authors Alliance filed an amicus brief, joined by EFF, ARL, ALA, and Public Knowledge, with the 9th Circuit in
In an earlier post, we shared details from Judge Alsup’s decision on Anthropic’s motion for summary judgment in Bartz v.
Late last week Judge Alsup, presiding over the Bartz v. Anthropic copyright AI litigation, granted a motion to certify a class representing authors and rightsholders of nearly 7 million books. If you are a book author (or a publisher, or an heir to an author), you should be paying attention because there is a good chance that you could be included in this class.
NO FAKES 2025 does not care about actual deception, impersonation, and harm to the average person; instead, it focuses on enabling political censorship and monetization of celebrity likeness.
“Market dilution” suggests that “using copyrighted books to train an LLM might harm the market for those works because it enables the rapid generation of countless works that compete with the originals, even if those works aren’t themselves infringing.”