To skirt copyright law, Anthropic bought and destroyed millions of physical books to train its "Claude" LLM
To skirt copyright law, Anthropic bought and destroyed millions of physical books to train its "Claude" LLM

Anthropic destroyed millions of print books to build its AI models

To comply with copyright law, not to skirt it. That's what companies that scan large numbers of books do. See for example Authors Guild v. Google from back when Google was scanning books to add to their book search engine. Framing this like it's some kind of nefarious act is misleading.
They also weren't destroying rare books. They were buying in-print books from major retailers, which means that while yes, that is environmentally wasteful, it's not actually destroying books in the classical destruction of knowledge sense since the manufacturer will just print another one if there's demand for it.
This as well. Growing up in a house of book lovers, myself included, destroying a book was akin to kicking a puppy. Realistically though, they're ultimately consumables. They're meant to be bought, used, and replaced as needed. With luck the destruction included recycling as much as possible, seeing as it's mainly paper.
Yeah, this is on the way of being a win. In this case they actually bought the books, which has been one of the biggest issues with LLMs. There's certainly more discussion to be had around how they use the materials in the end, but this is a step in the right direction.
To a certain extent I agree, but you can buy a book and still commit copyright infringement by copying its contents (for use other than personal use)
If this would go to court, it would depend on whether training an LLM model is more akin to copying or learning. I can see arguments for either interpretation, but I suspect that the law would lean more toward it being copying rather than learning