> Alsup ruled that Anthropic's use of copyrighted books to train its AI models was "exceedingly transformative" and qualified as fair use
> "All Anthropic did was replace the print copies it had purchased for its central library with more convenient space-saving and searchable digital copies for its central library — without adding new copies, creating new works, or redistributing existing copies"
It was always somewhat obvious that pirating a library would be copyright infringement. The interesting findings here are that scanning and digitizing a library for internal use is OK, and using it to train models is fair use.
> But Alsup drew a firm line when it came to piracy.
> "Anthropic had no entitlement to use pirated copies for its central library," Alsup wrote. "Creating a permanent, general-purpose library was not itself a fair use excusing Anthropic's piracy."
That is, he ruled that
- buying, physically cutting up, physically digitizing books, and using them for training is fair use
- pirating the books for their digital library is not fair use.
So Suno would only really need to buy the physical albums and rip them to be able to generate music at an industrial scale?
If the output from said model uses the voice of another person, for example, we already have a legal framework in place for determining if it is infringing on their rights, independent of AI.
Courts have heard cases of individual artists copying melodies, because melodies themselves are copyrightable: https://www.hypebot.com/hypebot/2020/02/every-possible-melod...
Copyright law is a lot more nuanced than anyone seems to have the attention span for.
But Suno is definitely not training models in their basement for fun.
They are a private company selling music, using music made by humans to train their models, to replace human musicians and artists.
We'll see what the courts say but that doesn't sound like fair use.
The law doesn't distinguish between basement and cloud – it's a service. You can sell access to the service without selling songs to consumers.
Suno can’t prevent humans from copying other humans, it can only make sure that the direct output of its system isn’t infringing.
In this context, this would be the equivalent of Suno explicitly placing stop points throughout the training, tokenization, and generation processes to verify that there was absolutely no chance of it generating copyrighted material through some kind of clean room reconstruction test. They would also need those tests to be audited at random by a third party governing body. Obviously they are not doing this, so the metaphor definitely does not track here.
Anything remotely beyond that and we have teams of humans adjudicating specific cases: https://library.mi.edu/musiccopyright/currentcases
Surely, for Suno to claim fair usage and be given free reign to build a commercial business off of literally anyone's original works then the bare minimum bar for allowing that usage would be: make a satisfactory test to prove that you're always doing something transformative and original, within practical limits.
But i guess I'm not surprised that 2025 has little respect for artists.
When a general computer using agent recreates songs in Logic Pro in high fidelity, then what?
It’s called Fair Use for a reason – we let humans Use things generally and ask them to be Fair.
Or we can go in the direction of movies and TV where screenshots of protected content show up blank on my iPhone. Just in case someone wanted to, god forbid, clip the show.
>It’s called Fair Use for a reason – we let humans Use things generally and ask them to be Fair.
So exhausted with people who come to these threads and try to discuss legal issues by only paying lip service to the words and not their meanings, let alone the actual law that they seem to want to debate. Then they go even further and turn it into some grand political statement, or hypothesize why copyright shouldn't exist at all. But there is absolutely no jurisprudence that would indicate a DAW is the kind of tool I described. I understand you came up with an argument in your head why it could be, but I'm letting you know that in the law, it's not what would be considered a reasonable argument and it would go nowhere.
DAWs are tools made to create music, generally. They do not contain banks of copyrighted materials to which the user ultimately pulls the copying "trigger" (that's the system I described).
I hope that helps.
Someone responded and said "Why not DAWs, then?" The answer is because a DAW is not that kind of service or machine.
>t’s easy to fall back to known concepts to frame new things, but that is not accurate. LLMs do not hold a “banks of copyrighted materials”,
As an aside. That's clearly not true in some models given that in a number of the cases, the plaintiffs can recreate their works verbatim.
DAWs are tools made to create music, generally. They do not contain banks of copyrighted materials to which the user ultimately pulls the copying "trigger" (that's the system I described).
You are quite literally describing sample packs (which are copyrighted). The only difference is that they figured out a fair licensing scheme for those. Is my understanding of copyright law wrong or poor here?Imagine we invented some new hypothetical technology to take all of the sample packs in the world as input and produce new sample packs that humans haven't thought of before. Should we figure out how to license those packs fairly or pretend we never invented it?
Only so many artists have the patience to make each drum from scratch.
Where are you on the continuum? Regarding training an AI model in my basement on purchased music, do you think I should:
- Not be allowed to train it
- Not be allowed to run it
- Not be allowed to share outputs from it anywhere
- Not be allowed to share outputs from it publicly
- Not be allowed to share outputs from it commercially
- Not be allowed to share its weights for others to run it
Or are you primarily focused on the current legal precedent?
Sure, I appreciate that. My point is that none of this has anything to do with § 1201 so there's really no point in coming to this with a kind of incredulity that is counterproductive stemming from your own beliefs about that one particular law. Not saying that is necessarily what you are doing, but I see that kind of approach so frequently here. A lot of not really knowing what a copyright protects, its limits, how they are adjudicated, etc, but then a lot of confidence about how it is all just wrong for society.
For starters, to answer your first question. Copyright protects creative artistic expressions. What is covered is defined in the copyright statute, and the list does not include massages. So, that would be the reason why a massage is not protected. Why is "massage" on that list? Probably because no one can reasonably consider a massage a creative artistic expression. Choreography is the art in which that form of expression exists and would be covered. Could you copyright a dance that included massage movements? Yeah, sure. Could you copyright a dance that consisted entirely of massage movements. Sure. Could you use that copyright to prevent massage therapists from "performing" massages? No.
That's obviously a very surface level take and what is actually protected in a copyright isn't necessarily the entirety of the work but the aspects of it that original expressions. There are other limitations too, like something being de minimis. You can't copyright "the sky was blue" (Scarlett Begonias, the Grateful Dead) and actually prohibit others from using the phrase. That phrase alone is too small (among other things). The Grateful Dead do have a copyright to the entirety of the lyrics to Scarlett Begonias and can control various kinds of uses of the the lyrics.
>Or are you primarily focused on the current legal precedent?
All litigators are focused on current legal precedent. You cannot make arguments for how things should be without regard for how things are as that is the fundamental basis for what should be changed and why.
>Where are you on the continuum? Regarding training an AI model in my basement on purchased music, do you think I should:
Personally, I find AI abhorrent. I think its wrong for it to be trained without any compensation to the authors of the works used in the training, and I think it's wrong for the output to be commercialized to the benefit of the owner of the model without any compensation to the authors of the works used in generating the outputs.
The problem is, copyright law wasn't written for machines. It was written for humans who create things.
In the case of songs (or books, paintings, etc), only humans and companies can legally own copyright, a machine can't. If an AI-powered tool generates a song, there’s no author in the legal sense, unless the person using the tool claims authorship by saying they operated the tool.
So we're stuck in a grey zone: the input is human, the output is AI generated, and the law doesn't know what to do with that.
For me the real debate is: Do we need new rules for non-human creation?
when you buy a book, you are not acceding to a license to only ever read it with human eyes, forbearing to memorize it, never to quote it, never to be inspired by it.
> Interestingly, Llama 1 65B, a similar-sized model released in February 2023, had memorized only 4.4 percent of Harry Potter and the Sorcerer's Stone. This suggests that despite the potential legal liability, Meta did not do much to prevent memorization as it trained Llama 3. At least for this book, the problem got much worse between Llama 1 and Llama 3.
> Harry Potter and the Sorcerer's Stone was one of dozens of books tested by the researchers. They found that Llama 3.1 70B was far more likely to reproduce popular books—such as The Hobbit and George Orwell’s 1984—than obscure ones. And for most books, Llama 3.1 70B memorized more than any of the other models.
Memorising isn't wrong but when machines memorise at scale and the people behind the original work get nothing, it raises big ethical questions.
The law hasn't caught up.
I also play the guitar, and it took me 10 years to learn 30 or 40 songs. So I don't see how anyone can learn 7 million songs in a couple of minutes.
Most AI seems much better at reproducing a semi-identical copies of an original work than existing video/audio encoders.
Now, what if instead of training myself using real instruments, I train my AI and do the same. Is it different?
It is complicated, but there are many arguments in favor of fair use, probably more than they are against but as you say, let's the courts decide.
But in any case, piracy is illegal in every case. As a human, it is illegal for me to use pirate copies, whether it is for training myself as a musician, for training my AI, or for simply listening.
Claude has been considered transformative given it's not really meant to generate books but Suno or Midjourney are absolutely in another category.
But as long as the model isnt outputting infringing works theres not really any issue there either.
Anthropic doesn't take books and use them to train a model that is intended to generate new books. (Perhaps it could do that, to some extent, but that's no its [sole] purpose.)
But Suno would be taking music to train a model in order to generate new music. Is that transformative enough? We don't know what a judge thinks, at least not yet.
If you read the ruling, training was considered fair use in part because Claude is not a book generation tool. Hence it was deemed transformative. Definitely not what Suno and Udio are doing.
Do keep in mind though: this is only for the wealthy. They're still going to send the Pinkertons at your house if you dare copy a Blu-ray.
Hey woah now, that's a Hasbro play, not a Disney one.
Found it: https://www.nbcnews.com/tech/tech-news/federal-judge-rules-c...
> “That Anthropic later bought a copy of a book it earlier stole off the internet will not absolve it of liability for the theft,” [Judge] Alsup wrote, “but it may affect the extent of statutory damages.”
In the UK it's a criminal offense if you distribute a copyrighted work with the intent to make gain or with the expectation that the owner will make a loss.
Gain and loss are only financial in this context.
Meaning that in both countries the copyright owner can sue you for copyright infringement.
Do you mean:
A) It's not a criminal offence?
B) The copyright owner cannot file a civil suit for damages?
C) Something else?
We will have a trial on the pirated copies used to create Anthropic’s central library and the resulting damages, actual or statutory (including for willfulness).
I think the distinction between civil and criminal trials is smaller in my home country. The fact that there is a trial at all implies that someone commited a ‘crime’.
Rhetorical question, we all know that me reading books is not "transformative" so it won't be considered fair use for me to yoink them (transformative as in transforming more damage to the society at large into more money for the already rich).
They simply don't want to and think they can skirt the law while the judges catch up.
If the book is out of print, then tough luck. That's not a license to infringe on the publisher's copyright. If we're not ok with that, we have legislative means to change that. A judge shouldn't be rewriting law in that manner.
But if you tried to open a black market selling that media: you'd be hunted down to the ends of the earth. Or to China/North Korea, at least.
Why would you ever do that? Nobody would buy it. They'd just get it in the same place you did.
Everyone can out a disk in a DVD player; sailing the high seas is much trickier.
Having access to a camera doesn't permit you to take the footage home to review.The company still owns that footage, after all.
Now, if you had your own camera recording everything at your desk... I guess that falls into one or two party states.
This conversation becomes incredibly unenjoyable when you pull rhetorical techniques like completely ignoring the entirety of what I wrote.
No it's not. And you ever heard of a publishing house? They don't need to negotiate with every single author individually. That's preposterous.
Yeah they do. What do you think the employees of a publishing house do? They make deals, work with authors, and accept/reject pitches. They 100% need to make sure every work is under a negotiated contract.
It's not the only reason fair use exists, but it's the thing that allows e.g. search engines to exist, and that seems pretty important.
> And you ever heard of a publishing house? They don't need to negotiate with every single author individually. That's preposterous.
There are thousands of publishing houses and millions of self-published authors on top of that. Many books are also out of print or have unclear rights ownership.
No, it kinda isn't. Show me anything that supports this idea beyond your own immediate conjecture right now.
>It's not the only reason fair use exists, but it's the thing that allows e.g. search engines to exist, and that seems pretty important.
No, that's the transformative element of what a search engine provides. Search engines are not legal because they can't contact each licensor, they are legal because they are considered hugely transformative features.
>There are thousands of publishing houses and millions of self-published authors on top of that. Many books are also out of print or have unclear rights ownership.
Okay, and? How many customers does Microsoft bill on a monthly basis?
It's inherent in the nature of the test. The most important fair use factor is the effect on the market for the work, so if the use would be uneconomical without fair use then the effect on the market is negligible because the alternative would be that the use doesn't happen rather than that the author gets paid for it.
> No, that's the transformative element of what a search engine provides. Search engines are not legal because they can't contact each licensor, they are legal because they are considered hugely transformative features.
To make a search engine you have to do two things. One is to download a copy of the whole internet, the other is to create a search index. I'm talking about the first one, you're talking about the second one.
> Okay, and? How many customers does Microsoft bill on a monthly basis?
Microsoft does this with an automated system. There is no single automated system where you can get every book ever written, and separately interfacing with all of the many systems needed in order to do it is the source of the overhead.
If your business model is not economically sustainable in the current legal landscape you operate in, the correct outcome is you go out of business.
There's lots and lots of potential businesses, infinite in fact, that fall into this understanding. They don't exist because they can't because we don't want them to, so you never see them. Which might give the impression of a right to scale, but no, it does not exist.
No, that's not the most important factor. The transformative factor is the most important. Effect on market for the work doesn't even support your argument anyway. Your argument is about the cost of making the end product, which is totally distinct from the market effects on the copyright holder when the infringer makes and releases the infringing product.
>To make a search engine you have to do two things. One is to download a copy of the whole internet, the other is to create a search index. I'm talking about the first one, you're talking about the second one.
So? That doesn't make you right. Go read the opinions, dude. This isn't something that's actually up for debate. Search engines are fair uses because of their transformative effect, not because they are really expensive otherwise. Your argument doesn't even make sense. By that logic, anything that's expensive becomes a fair use. It's facially ridiculous. Them being expensive is neither sufficient nor necessary for them to be a fair use. Their transformative nature is both sufficient and necessary to be found a fair use. Full stop.
>Microsoft does this with an automated system. There is no single automated system where you can get every book ever written, and separately interfacing with all of the many systems needed in order to do it is the source of the overhead.
Okay, and? They don't need to get every single book ever written. The libraries they pirated do not consist of "every single book ever written". It's hard to take this argument in good faith because you're being so ridiculous.
It's a four factor test because all of the factors are relevant, but if the use has negligible effect on the market for the work then it's pretty hard to get anywhere with the others. For example, for cases like classroom use, even making verbatim copies of the entire work is often still fair use. Buying a separate copy for each student to use for only a few minutes would make that use uneconomical.
> Effect on market for the work doesn't even support your argument anyway. You're argument is about the cost of making the end product, which is totally distinct from the market effects on the copyright holder when the infringer makes and releases the infringing product.
We're talking about the temporary copies they make during training. Those aren't being distributed to anyone else.
> So? That doesn't make you right.
Making a copy of everything on the internet is a prerequisite to making a search engine. It's something you have to do as a step to making the index, which is the transformative step. Are you suggesting that doing the first step is illegal or what do you propose justifies it?
> By that logic, anything that's expensive becomes a fair use. It's facially ridiculous.
Anything with unreasonably high transaction costs. Why is that ridiculous? It doesn't exempt any of the normal stuff like an individual person buying an individual book.
> They don't need to get every single book ever written.
They need to get as many books as possible, with the platonic ideal being every book. Whether or not the ideal is feasible in practice, the question is whether it's socially beneficial to impose a situation with excessively high transaction costs in order to require something with only trivial benefit to authors (potentially selling one extra copy).
All four factors are not equally relevant which is something described in pretty much every single fair use opinion. Educational uses are educational uses and considered fair because of their educational purpose (purpose is one of the factors), again, not because it's expensive. Maybe next time try googling or using ChatGPT "fair use educational".
>We're talking about the temporary copies they make during training. Those aren't being distributed to anyone else.
It's your argument. Not mine. You do not understand the market harm factor and it has nothing to do with Anthropic's transaction costs. That's just fully outright absolutely incorrect application of law.
>Making a copy of everything on the internet is a prerequisite to making a search engine. It's something you have to do as a step to making the index, which is the transformative step. Are you suggesting that doing the first step is illegal or what do you propose justifies it?
The transformative step is why it's a fair use, not the "market harm" (which you misunderstand) or the made up argument that it's "too expensive". In fact, I said this like every single turn in our conversation so it's a bit perplexing to me that you can now ask me "do you mean that it being transformative is what makes it legal" when that was my exact argument three times.
>Anything with unreasonably high transaction costs. Why is that ridiculous? It doesn't exempt any of the normal stuff like an individual person buying an individual book.
It's ridiculous because of the example I gave. Things being expensive is not a defense to copyright infringement and copyright law has no obligation to make expensive business models work. Copyright has an obligation to make transformative business models work because of the overall good they provide to society. Describing it as a "transaction cost" just kicks the can down the road even further and doesn't deal with the substance, either. They could have gone to the major publishers and licensed books from them. They didn't. That's generally who they are being sued by. When they are being sued by copyright owners in the fringe examples you pointed to, they will become relevant then.
>They need to get as many books as possible, with the platonic ideal being every book. Whether or not the ideal is feasible in practice, the question is whether it's socially beneficial to impose a situation with excessively high transaction costs in order to require something with only trivial benefit to authors (potentially selling one extra copy).
Lol dude, it was your example, not mine. They do not need every single book. They aren't being sued over every single book anyway, so it's totally besides the point.
Substitute infringement for theft.
No, that doesn't undo the infringement. At most, that would mitigate actual damages, but actual damages aren't likely to be important, given that statutory damages are an alternative and are likely to dwarf actual damages. (It may also figure into how the court assigns statutory damages within the very large range available for those, but that range does not go down to $0.)
> They will have ceased and desisted.
"Cease and desist" is just to stop incurring additional liability. (A potential plaintiff may accept that as sufficient to not sue if a request is made and the potential defendant complies, because litigation is uncertain and expensive. But "cease and desist" doesn't undo wrongs and neutralize liability when they've already been sued over.)
For anyone else who wants to do the same thing though this is likely all they need to do.
Cutting up and scanning books is hard work and actually doing the same thing digitally to ebooks isn't labor free either, especially when they have to be downloaded from random sites and cleaned from different formats. Torrenting a bunch of epubs and paying for individual books is probably cheaper
Setting the penalty to what it would have cost to obey the law in the first place does the opposite.
If you give people a claim for damages which is an order of magnitude larger than their actual damages, it encourages litigiousness and becomes a vector for shakedowns because the excessive cost of losing pressures innocent defendants to settle even if there was a 90% chance they would have won.
Meanwhile both parties have the incentive to settle in civil cases when it's obvious who is going to win, because a settlement to pay the damages is cheaper than the cost of going to court and then having to pay the same damages anyway. Which also provides a deterrent to doing it to begin with, because even having to pay lawyers to negotiate a settlement is a cost you don't want to pay when it's clear that what you're doing is going to have that result.
And when the result isn't clear, penalizing the defendant in a case of first impression isn't just either, because it wasn't clear and punitive measures should be reserved for instances of unambiguous wrongdoing.
> - buying, physically cutting up, physically digitizing books, and using them for training is fair use
> - pirating the books for their digital library is not fair use.
That seems inconsistent with one another. If it's fair use, how is it piracy?
It also seems pragmatically trash. It doesn't do the authors any good for the AI company to buy one copy of their book (and a used one at that), but it does make it much harder for smaller companies to compete with megacorps for AI stuff, so it's basically the stupidest of the plausible outcomes.
* They downloaded a massive online library of pirated books that someone else was distributing illegally. This was not fair use.
* They then digitised a bunch of books that they physically owned copies of. This was fair use.
This part of the ruling is pretty much existing law. If you have a physical book (or own a digital copy of a book), you can largely do what you like with it within the confines of your own home, including digitising it. But you are not allowed to distribute those digital copies to others, nor are you allowed to download other people's digital copies that you don't own the rights to.
The interesting part of this ruling is that once Anthropic had a legal digital copy of the books, they could use it for training their AI models and then release the AI models. According to the judge, this counts as fair use (assuming the digital copies were legally sourced).
Can you point me to the US Supreme Court case where this is existing law?
It's pretty clear that if you have a physical copy of a book, you can lend it to someone. It also seems pretty reasonable that the person borrowing it could make fair use of it, e.g. if you borrow a book from the library to write a book review and then quote an excerpt from it. So the only thing that's left is, what if you do the same thing over the internet?
Shouldn't we be able to distinguish this from the case where someone is distributing multiple copies of a work without authorization and the recipients are each making and keeping permanent copies of it?
Also, I don't quite understand how your example is relevant to the case. If you give a book to a friend, they are now the owner of that book and can do what they like with it. If you photocopy that book and give them the photocopy, they are not the owner of the book and you have reproduced it without permission. The same is, I believe, true of digital copies - this is how ebook libraries work.
In this case, Anthropic were the legal owners of the physical books, and so could do what they wanted with them. They were not the legal owners of the digital books, which means they can get prosecuted for copyright infringement.
We're talking about lending rather than ownership transfers, though of course you could regard lending as a sort of ownership transfer with an agreement to transfer it back later.
> If you photocopy that book and give them the photocopy, they are not the owner of the book and you have reproduced it without permission.
But then the question is whether the copy is fair use, not who the owner of the original copy was, right? For example, you can make a fair use photocopy of a page from a library book.
> They were not the legal owners of the digital books, which means they can get prosecuted for copyright infringement.
Even if the copy they make falls under fair use and the person who does own that copy of the book has no objection to their doing this?
If you photocopy a single page from a library book, this is often (but not always) fair use because you're copying only a limited part of the book. In the same way, you can quote a section or paragraph of a book under fair use. You cannot copy the whole book, though. Therefore:
> Even if the copy they make falls under fair use and the person who does own that copy of the book has no objection to their doing this?
If the copy had been made under fair use, then yes, this wouldn't be illegal. But it wasn't, because it was a reproduction and distribution of the entire book by someone who did not have the right to do that.
I understand that some of these things might be confusing to you, but Anthropic is absolutely within the position of being able to afford attorneys and get good advice as to what they could legally. I hope you also understand that good legal advice isn't being told what you want so you can do the thing you want to do without any regard for what are likely outcomes.
With that in mind, what do you think the inconsistency is between ReDigi and Sony?
He said:
> It was always somewhat obvious that pirating a library would be copyright infringement.
??
> pirating the books for their digital library is not fair use.
"Pirating" is a fuzzy word and has no real meaning. Specifically, I think this is the cruz:
> without adding new copies, creating new works, or redistributing existing copies
Essentially: downloading is fine, sharing/uploading up is not. Which makes sense. The assertion here is that Anthropic (from this line) did not distribute the files they downloaded.
It's a bit surprising that you can suddenly download copyrighted materials for personal use and and it's kosher as long as you don't share them with others.
I never saw any of these. All the cases I saw were related to people using torrents or other P2P software (which aren't just downloading). These might exist, but I haven't seen them.
> It's a bit surprising that you can suddenly download copyrighted materials for personal use and it's kosher as long as you don't share them with others.
Every click on a link is a risk of downloading copyrighted material you don't have the rights to.
Searching the internet, it appears that it's a civil infraction, but it's also confused with the notion that "piracy" is illegal, a term that's used for many different purposes. I see "It is illegal to download any music or movies that are copyrighted." under legal advice, which I know as a statement is not true.
Hence my confusion.
I should note: I'm not arguing from the perspective of whether it's morally or ethically right. Only that even in the context of this thread, things are phrased that aren't clear.
Gonzalez is a ruling about downloading even though there was also distribution.
After all, illegally downloading research papers in order to write new ones is highly transformative.
I've fixed your question so that it accurately represents what I said and doesn't put words in my mouth.
If I click on a link and download a document, is that illegal?
I do not know if the person has the right to distribute it or not. IANAL, but when people were getting sued by the RIAA years back, it was never about downloading, but also distribution.
As I said, IANAL, but feel free to correct me, but my understanding is that downloading a document from the internet is not illegal.
Did you mean to write "but about distribution" here?
That suggests otherwise.
https://www.courtlistener.com/docket/67569326/598/kadrey-v-m...
Note: I am not a lawyer.
https://investors.autodesk.com/news-releases/news-release-de...
What you’re saying is like calling Al Capone a tax cheat. Nonsense.
They went after Aaron over copyright.
Here's an article explaining in more detail [1].
Most experts say that if Swartz had gone to trial and the prosecution had proved everything they alleged and the judge had decided to make an example of Swartz and sentence harshly it would have been around 7 years.
Swartz's own attorney said that if they had gone to trail and lost he thought it was unlikely that Swartz would get any jail time.
Swartz also had at least two plea bargain offers available. One was for a guilty plea and 4 months. The other was for a guilty plea and the prosecutors would ask for 6 months but Swartz could ask the judge for less or for probation instead and the judge would pick.
[1] https://www.popehat.com/2013/02/05/crime-whale-sushi-sentenc...
When I clicked the link, I got an article about a business that was selling millions of dollars of pirated software.
This guy made millions of dollars in profit by selling pirated software. This wasn't a case of transformative works, nor of an individual doing something for themselves. He was plainly stealing and reselling something.
Seven years for thumbing your nose at Autodesk when armed robbery would get you less time says some interesting things about the state of legal practice.
This absolutely falls under copyright law as I understand it (not a lawyer). E.g. the disclaimer that rolls before every NFL broadcast. The notice states that the broadcast is copyrighted and any unauthorized use, including pictures, descriptions, or accounts of the game, is prohibited. There is wiggle room for fair use by news organizations, critics, artists, etc.
Make no mistake, they’re seeking to exploit the contents of that material for profits that are orders of magnitude larger than what any shady pirated-material reseller would make. The world looks the other way because these companies are “visionary” and “transformational.”
Maybe they are, and maybe they should even have a right to these buried works, but what gives them the right to rip up the rule book and (in all likelihood) suffer no repercussions in an act tantamount to grand theft?
There’s certainly an argument to be had about whether this form of research and training is a moral good and beneficial to society. My first impression is that the companies are too opaque in how they use and retain these files, albeit for some legitimate reasons, but nevertheless the archival achievements are hidden from the public, so all that’s left is profit for the company on the backs of all these other authors.
This is very different to what Anthropic did. Nobody was buying copies of books from Anthropic instead of the copyright holder.
We've only dealt with the fairly straight-forward legal questions so far. This legal battle is still far from being settled.
There's already one decision on a competitor.
It makes sense, if you think of how the model works.
Not even the authors suing Anthropic have claimed it can do this, have they?
As an extreme example, consider murder. Obviously it should be illegal, but if it's legal for one group and not for another, the group for which it's illegal will probably be wiped out, having lost the ability to avenge deaths in the group.
It's much more important that laws are applied impartially and equally than that they are even a tiny bit reasonable.
Laws and their enforcement are a clusterfuck. To achieve greater justice we should strive towards better judgements overall.
God, stop with the group on group bs please and engage with things the way they're written without injecting the entirety of your cynical worldview layered on top.
Furthermore, group affiliation based differences in judicial decisions are very common, both when it comes to ethnic origin, wealth and profession.
In this case group affiliation is also directly relevant: individuals who have infringed copyright are typically not treated in the way that these firms that have infringed copyright are. The group affiliation in question is thus 'are you an employer/wealth person owning part of a large firm' vs 'a normal, non-employer/non-wealthy person'.
If there are precedents with a certain application, then they must continue or be overturned generally.
Correct but uneven application of a law is more dangerous than incorrect but even application.
in other words, provided you have enough spare capital to spin up a corporation, you can break the law!!!!
This is reaching at best.
"Pirates" also transform the works they distribute. They crack it, translate it, compress it to decrease download times, remove unnecessary things, make it easier to download by splitting it in chunks (essential with dial-up, less so nowadays), change distribution formats, offer it trough different channels, bundle extra software and media that they themselves might have coded like trainers, installers, sick chiptunes and so on. Why is the "transformation" done by a big corpo more legal in your views?
your piracy apologetics are obviously irrelevant.
Come up with a better comparison.
First, Authors argue that using works to train Claude’s underlying LLMs
was like using works to train any person to read and write, so Authors
should be able to exclude Anthropic from this use (Opp. 16).
Second, to that last point, Authors further argue that the training was
intended to memorize their works’ creative elements — not just their
works’ non-protectable ones (Opp. 17).
Third, Authors next argue that computers nonetheless should not be
allowed to do what people do.
https://media.npr.org/assets/artslife/arts/2025/order.pdf> First, Authors argue that using works to train Claude’s underlying LLMs was like using works to train any person to read and write, so Authors should be able to exclude Anthropic from this use (Opp. 16). But Authors cannot rightly exclude anyone from using their works for training or learning as such. Everyone reads texts, too, then writes new texts. They may need to pay for getting their hands on a text in the first instance. But to make anyone pay specifically for the use of a book each time they read it, each time they recall it from memory, each time they later draw upon it when writing new things in new ways would be unthinkable. For centuries, we have read and re-read books. We have admired, memorized, and internalized their sweeping themes, their substantive points, and their stylistic solutions to recurring writing problems.
Couldn't have put it better myself (though $deity knows I tried many times on HN). Glad to see Judge Alsup continues to be the voice of common sense in legal matters around technology.
Yep, that name's a blast from the past! He was the judge on the big Google/Oracle case about Android and Java years ago, IIRC. I think he even learned to write some Java so he could better understand the case.
“We have admired, memorized, and internalized their sweeping themes, their substantive points, and their stylistic solutions to recurring writing problems.”
Claude is not doing any of these things. There is no admiration, no internalizing of sweeping themes. There’s a network encoding data.
We’re talking about a machine that accepts content and then produces more content. It’s not a person, it’s owned by a corporation that earns money on literally every word this machine produces. If it didn’t have this large corpus of input data (copyrighted works) it could not produce the output data for which people are willing to pay money. This all happens at a scale no individual could achieve because, as we know, it is a machine.
That the mechanism performing these things is a network encoding data is… well, that description, at that level of abstraction, is a similarity with the way a human does it, not even a difference.
My network is a 3D mess made of pointy bi-lipid bags exchanging protons across gaps moderated by the presence of neurochemicals, rather than flat sheets of silicon exchanging electrons across tuned energy band-gaps moderated by other electrons, but it's still a network.
> We’re talking about a machine that accepts content and then produces more content. It’s not a person, it’s owned by a corporation that earns money on literally every word this machine produces. If it didn’t have this large corpus of input data (copyrighted works) it could not produce the output data for which people are willing to pay money. This all happens at a scale no individual could achieve because, as we know, it is a machine.
My brain is a machine that accepts content in the form of job offers and JIRA tickets (amongst other things), and then produces more content in the form of pull requests (amongst other things). For the sake specifically of this question, do the other things make a difference? While I count as a person and am not owned by any corporation, when I work for one, they do earn money on the words this biological machine produces. (And given all the models which are free to use, the LLMs definitely don't earn money on "literally" every word those models produce). If I didn't have the large corpus of input data — and there absolutely was copyright on a lot of the school textbooks and the TV broadcast educational content of the 80s and 90s when I was at school, and the Java programming language that formed the backbone of my university degree — I could not produce the output data for which people are willing to pay money.
Should corporations who hire me be required to pay Oracle every time I remember and use a solution that I learned from a Java course, even when I'm not writing Java?
That the LLMs do this at a scale no individual could achieve because it is a machine, means it's got the potential to wipe me out economically. Economics threat of automation has been a real issue at least since the luddites if not earlier, and I don't know how the dice will fall this time around, so even though I have one layer of backup plan, I am well aware it may not work, and if it doesn't then government action will have to happen because a lot of other people will be in trouble before trouble gets to me (and recent history shows that this doesn't mean "there won't be trouble").
Copyright law is one example of government action. So is mandatory education. So is UBI, but so too is feudalism.
Good luck to us all.
It's not an issue because it's not currently illegal because nobody could have foreseen this years ago.
But it is profiting off of the unpaid work of millions. And there's very little chance of change because it's so hard to pass new protection laws when you're not Disney.
The model is very different.
This argument is more along the lines of: blaming Microsoft Word for someone typing characters into the word processors algorithm, and outputting a copy of an existing book. (Yes, it is a lot easier, but the rationale is the same). In my mind the end user prompting the model would be the one potentially infringing.
I do think that a big part of the reason Anthropic downloaded millions of books from pirate torrents was because they needed that input data in order to generate the output, their product.
I don’t know what that is, but, IMHO, not sharing those dollars with the creators of the content is clearly wrong.
I don't think humans learn via backprop or in rounds/batches, our learning is more "online".
If I input text into an LLM it doesn't learn from that unless the creators consciously include that data in the next round of teaching their model.
Humans also don't require samples of every text in history to learn to read and write well.
Hunter S Thompson didn't need to ingest the Harry Potter books to write.
“Piracy” is mostly a rhetorical term in the context of copyright. Legally, it’s still called infringement or unauthorized copying. But industries and lobbying groups (e.g., RIAA, MPAA) have favored “piracy” for its emotional weight.
Copyright infringement is unauthorized reproduction - you have made a copy of something, but you have not deprived the original owner of it. At most, you denied them revenue although generally less than the offended party claims, since not all instances of copying would have otherwise resulted in a sale.
And to be clear, we javelin the word infringement precisely because it is not theft.
In addition to the deprived revenue, piracy also improves on the general relevance the author has or may have in the public sphere. Essentially, one of the side effects of piracy is basically advertising.
Doctorow was one of the early ones to bring this aspect of it up.
Real piracy always involves booty.
Naturally booty is wealth that has been hoarded.
Has nothing to do with wealth that may or may not come in the future, regardless of whether any losses due to piracy have taken place already or not.
Referring to this? (Wikipedia's disambiguation page doesn't seem to have a more likely article.)
https://en.wikipedia.org/wiki/Richard_Stallman#Copyright_red...
Stallman places great importance on the words and labels people use to talk about the world, including the relationship between software and freedom. He asks people to say free software and GNU/Linux, and to avoid the terms intellectual property and piracy (in relation to copying not approved by the publisher). One of his criteria for giving an interview to a journalist is that the journalist agrees to use his terminology throughout the article.
(As an aside, it seems pointless to decry it as a "talking point". The reason it was brought up is presumably because the author agrees with it and thinks it's relevant. It's also entirely possible that the author, like me, made this argument without being aware that it was popularized by Richard Stallman. If it makes sense then you can hear the argument without hearing the person and still find it agreeable.)
"Piracy" is used to refer to copyright violation to make it sound scary and dangerous to people who don't know better or otherwise don't think about it too hard. Just imagine if they called it "banditry" instead; now tell me that pirates are not bandits with boats. They may as well have called it banditry and it's worth correcting that. (I also think it's worth ridiculing but that doesn't appear to be Stallman's primary point.) It's not banditry (how ridiculous would it be to call it that?), it's copyright infringement.
Edit:
Reading my comment again in the context of other things you wrote, I suspect the argument will not pass muster because you do not seem to see piracy's change in meaning as manufactured by PR work purchased by media industry leaders. I'm not really trying to convince you that it's true but it may be worth considering that it is the fundamental disagreement you seem to have with others on Stallman's point; again, not saying you're wrong, just that's where the disagreement is.
In short the post is bait.
This is an uncharitable interpretation. The ostensible point of the comment, or at least a stronger and still-reasonable interpretation, is that they are trying to point out that this specific word choice confuses concepts, which it does. Richard Stallman and the commenter in question are absolutely correct to point that out. You actually seem to be agreeing with Stallman, at least in the abstract.
It's should be acknowledged how/why the meaning of the word changed. As I said, that seems to have been manufactured, which suggests, at least to me, that their (and Richard Stallman's) point is essentially the same as yours. That is to say, the US media industry started paying PR firms to use "piracy" as meaning something other than its normal definition until that became the common definition.
They should not purposely use a different definition like that. That is Stallman's point, and why he refuses to say "piracy" instead of "copyright infringement"; ocean banditry is not copyright infringement and it is confusing -- intentionally so -- to say that it is.
https://torrentfreak.com/spotifys-beta-used-pirate-mp3-files...
Funky quote:
> Rumors that early versions of Spotify used ‘pirate’ MP3s have been floating around the Internet for years. People who had access to the service in the beginning later reported downloading tracks that contained ‘Scene’ labeling, tags, and formats, which are the tell-tale signs that content hadn’t been obtained officially.
https://www.forbes.com/2009/08/04/online-anime-video-technol...
https://venturebeat.com/business/crunchyroll-for-pirated-ani...
My theory is that once they saw how much traffic they were getting, they realized how big of a market (subbed/dubbed) anime was.
- riding a wave of change
- not caring too much about legal constraints (or like they would say now "distrupting" the market, which very very often means doing illigal shit which beings them far more money then any penalties they will ever face from it)
- or caring about ethics too much
- and for recent years (starting with Amazone) a lot of technically illegal financing (technically undercutting competitors prices long term based on money from else where (e.g. investors) is unfair competitive advantage (theoretically) clearly not allowed by anti monopoly laws. And before you often still had other monopoly issues (e.g. see wintel)
So yes not systematic not complying with law to get unfair competitive advantage knowing that many of the laws are on the larger picture toothless when applied to huge companies is bread and butter work of US tech giants
breaking this laws is what lifted them from big, to supper marked dominant to a point where they have monopoly like power
that is _never_ good for everyone, or even good for the majority long term
what is good for everyone (but a few rich people and sometimes the US government) is proper fair competition. It drives down prices and allows people to vote with their money, a it is a corner stone of the American dream it pushes innovation and makes sure a country isn't left behind. Monopoly like companies on the other hand tend to have exactly the other effect, higher prices (long term), corruption, stagnating innovation, and a completely shattered American sound pretty bad for the majority of Americans.
Are all music creators better off now than before Spotify?
And Spotify is a bad example as it ran into another psudo monopoly with very unreasonable/unhealthy power (the few large music labels holding rights to the majority of main stream music).
They pretty much forced very bad terms onto Spotify which is to some degree why Spotify is pushing podcasts, as they can't be long term profitable with Music (raising prices doesn't help if the issue is a percent cut which rises too :/ )
Society underestimates the chasm that exists between an idea and raising sufficient capital to act on those ideas.
Plenty of people have ideas.
We only really see those that successfully cross it.
Small things EULA breaches, consumer licenses being used commercially for example.
If you're an individual pirating software or media, then from the rights owners' perspective, the most rational thing to do is to make an example of you. It doesn't happen everyday, but it does happen and it can destroy lives.
If you're a corporation doing the same, the calculation is different. If you're small but growing, future revenues are worth more than the money that can be extracted out of you right now, so you might get a legal nastygram with an offer of a reasonable payment to bring you into compliance. And if you're already big enough to be scary, litigation might be just too expensive to the other side even if you answer the letter with "lol, get lost".
Even in the worst case - if Anthropic loses and the company is fined or even shuttered (unlikely) - the people who participated in it are not going to be personally liable and they've in all likelihood already profited immensely.
And the system set up by society doesn't truly account for this or care.
but systematic wide spread big things and often many of them, giving US giant a unfair combative advantage
and don't think if you are a EU company you can do the same in the US, nop nop
but naturally the US insist that US companies can do that in the EU and complain every time a US company is fined for not complying for EU law
The AI sector, famously known for its inability to raise funding. Anthropic has in the last four years raised 17 billion dollars
Other industries do not have it this easy.
This is a narrative that gets passed around in certain circles to justify stealing content.
In this context, stealing is often used as a pejorative term to make piracy sound worse than it is. Except for mass distribution, piracy is often regarded as a civil wrong, and not a crime.
It would be more clear if you stick to either legal or colloquial variants, instead of switching back and forth. (Tbf, the judge in this case also used the term “piracy” colloquially).
In this case it seemed like you were making a point about the strict legal sense of a word, but misusing a different legal term to do so.
Hypocrisy requires deception, not ignorance.
Pirating 7 million books, remixing their content, and using that to make money on Claude.ai is like counterfeiting 7 million branded products and selling them on your Shopify website. The original creators don't get payment, and someone's profiting off their work.
Try doing that yourself and you'd get a knock on the door real quick.
Also mostly this would be a civil lawsuit for "damages".
The trial is scheduled for December 2025. That’s when a jury will decide how much Anthropic owes for copying and storing over seven million pirated books
Now places like Flea markets have been known to have a counterfeit DVD or two.
And there is more than one way to compare to non-digital content.
Regular books and periodicals can be sold out and/or out-of-print, but digital versions do not have these same exact limitations.
A great deal of the time though, just the opposite occurs, and a surplus is printed that no one will ever read, and which will eventually be disposed of.
Newspapers are mainly in the extreme category where almost always a significant number of surplus copies are intentionally printed.
It's all part of the same publication, a huge portion of which no one has ever rightfully expected for every copy to earn anything at all, much less a return on every single copy making it back to the original creator.
Which is one reason why so much material is supported by ads. Even if you didn't pay a high enough price to cover the cost of printing, it was all paid for well before it got into your hands.
Digital copies which are going unread are something like that kind of surplus. If you save it from the bin you should be able to do whatever you want with it either way, scan it how you see fit.
You just can't say you wrote it. That's what copyright is supposed to be for.
Like at the flea market, when two different vendors are selling the same items but one has legitimately purchased them wholesale and the other vendor obtained theirs as the spoils of a stolen 18-wheeler.
How do you know which ones are the pirated items?
You can tell because the original owners of the pirated cargo suffered a definite loss, and have none of it remaining any more.
OTOH, with things like fake Nikes at the flea market, you can be confident they are counterfeit whether they were stolen from anybody in any way or not.
Don’t we already have laws covering this? For example, sometimes excess books can be thrown in the bin. Often, they have the covers removed. Some will say something to the effect that “if you’ve received this without a cover it is a copyright violation.” I think one of the points of the lawsuit is it gives copyright holders discretion as to how their works are used/sold etc. The idea that “if you saved it from the bin you can do with it whatever you want” strips them of that right.
You could split hairs over whether saving an item from the bin occurred after a procedure to remove covers and it was already dumped, or before any contemplation was made about if or when dumping would take place.
Saving either way would be preserving what would otherwise be lost, even if it was well premeditated in advance of any imminent risk.
What if it was the last remaining copy?
Or even the only copy ever in existence of an original manuscript?
It's just not a concept suitable for a black & white judgment.
That's a very good sign that probably an entire book of regulations needs to be thrown out instead, and a new law written to replace it with something more sensible.
Or even the only copy ever in existence of an original manuscript?
I think these still remove the copyright of the author. As it stands, I have the right to write the best novel about the human condition ever conceived and also the right (if copyrighted) to not allow anyone to read it. I can light it on fire if I wish. I am not obligated to sell it to anyone. In the context of the above, I can stipulate that nobody can distribute excess copies even if they would be otherwise destroyed. You may think that’s wasteful or irrational but we have all kinds of rights that protect our ability to do irrational things with our own property.
>That's a very good sign that probably an entire book of regulations needs to be thrown out instead, and a new law written to replace it with something more sensible.
This sentiment implies that you do not think the owner has those rights. That’s fine, but there are plenty of people (myself included) who think those are reasonable rights. Intellectual property clause is in the first article of the US Constitution for a good reason, although I do think it can be abused.
Don’t have legal access to training data? Simply steal it, but move fast enough to keep ahead of the law. By the time lawsuits hit the company is worth billions and the product is embedded in everyday life.
That's a statement carefully crafted to be impossible to disprove. Of course they shipped pirated music (I've seen the files). Of course anyone paying attention knew. Nothing in the music industry was "clean" in those days. But, sure, no credible evidence because any evidence anyone shows you you'll decide is not credible. It's not in anyone's interests to say anything and none of it matters.
https://www.computerworld.com/article/1447323/google-reporte...
Facebook's "pivot to video" similarly relied on user-uploaded unlicensed video content, now not just pulling from television and film, but from content creators on platforms like YouTube.
Today, every "social" platform is now littered with "no copyright infringement intended" and "all credit to the original" copy-and-paste junk. Don't get me wrong, I'm a fan of remix culture – but I believe appropriating and monetizing the work of others without sharing the reward is a destructive cycle. And while there are avenues for addressing this, they're designed for the likes of Universal, Sony, Disney, etc. (I've had original recordings of original music flagged by megacorps because the applause triggered ContentID.)
AI slop further poisons the well. It's rough going out there.
It's not a common business practice. That's why it's considered newsworthy.
People on the internet have forgotten that the news doesn't report everyday, normal, common things, or it would be nothing but a listing of people mowing their lawns or applying for business loans. The reason something is in the news is because it is unusual or remarkable.
"I saw it online, so it must happen all the time" is a dopy lack of logic that infects society.
Edit: Apologies, I can’t edit it anymore.
And the playstation classic used an opensource ps1 emulator.
There was also some steam game ported from GameCube, and it had the Dolphin Emulator FPS counter in the corner of part of the trailer :D
I also remember reading that 2 of the PCSX2 devs ended up working on the EmotionEngine chip emulator for PS3 consoles with partial software emulation of PS2 (The CECH 02 and later models where they removed the EmotionEngine chip)
What you really should be asking is whether they infringed on the copyrights of the rippers. /s
Daniel Ek said: "my mission is to make music accessible and legal to everyone, while ensuring artists and rights holders got paid"
Also, the Swedish government has zero tolerance for piracy.
Stealing is stealing. Let's stop with the double standards.
They make money off the model weights, which is fair use (as confirmed by recent case law).
If it outputs parts of the book verbatim then that's a different story.
Pirating 7 million books, remixing their content, and using that to power Claude.ai is like counterfeiting 7 million branded products and selling them on your personal website. The original creators don't get credit or payment, and someone’s profiting off their work.
All this happens while authors, many of them teachers, are left scratching their heads with four kids to feed
But it does...
Please keep in mind, copyright is intended as a compromise between benefit to society and to the individual.
A thought experiment, students pirating textbooks and applying that knowledge later on in their work?
Meanwhile other cases have been less friendly to it being fair use, AI companies are already paying vast sums to publishers who presumably they wouldn’t if they felt confident it was “the law”, and on and on.
I don’t like arguing from “it’s the law”. A lot of law is terrible. What’s right? It’s clear to me that if AI gets good enough, as it nearly is now, it sucks a lot of profit away from creators. That is unbalanced. The AI doesn’t exist without the creators, the creators need to exist for our society to be great (we want new creative works, more if anything). Law tends to start conservatively based on historical precedent, and when a new technology comes along it often errs on letting it do some damage to avoid setting a bad precedent. In time it catches up as society gets a better view of things.
The right thing is likely not to let our creative class be decimated so a few tech companies become fantastically wealthy - in the long run, it’s the right thing even for the techies.
In my opinion, it will be upheld.
Looking at what is stored and the manner which it is stored. It makes sense that it's fair use.
If by "what is stored and the manner which it is stored" is intended to signal model weights, I'm not sure what the argument is? The four factors of copyright in no way mention a storage medium for data, lossless or loss-y.
(1) the purpose and character of the use, including whether such use is of a commercial nature or is for nonprofit educational purposes; (2) the nature of the copyrighted work; (3) the amount and substantiality of the portion used in relation to the copyrighted work as a whole; and (4) the effect of the use upon the potential market for or value of the copyrighted work.
In my opinion, this will likely see a supreme court ruling by the end of the decade.
A trillion parameter SOTA model is not substantially comprised of the one copyrighted piece. (If it was a Harry Potter model trained only on Harry Potter books this would be a different story).
Embeddings are not copy paste.
The last point about market impact would be where they make their argument but it's tenuous. It's not the primary use of AI models and built in prompts try to avoid this, so it shouldn't be commonplace unless you're jail breaking the model, most folk aren't.
Easy for the pirate to say. Artists might argue their intent was to trade compensation for one's personal enjoyment of the work.
That article also focuses on larger media and "moderate" amounts of piracy, so there's absolutely caveats in your claim.
"As with other studies, Kim and his colleagues found that when enforcement is low and piracy is rampant, both manufacturers and retailers suffer."
“'The implication is simply that, situated in a real-world context, our manufacturer and retailer should recognize that a certain level of piracy or its threat might actually be beneficial'..."
that isn't "just" stealing, it's organized crime
I get the sentiment, but that statement as is, is absurdly reductive. Details matter. Even if someone takes merchandise from a store without paying, their sentence will vary depending on the details.
Yes, but copying isn't stealing, because the person you "take" from still has their copy.
If you're allowed to call copying stealing, then I should be allowed to call hysterical copyright rabblerousing rape. Quit being a rapist, pyman.
There's so many texts, and they're so sparse that if I could copyright a work and never publish it, the restriction would be irrelevant. The probability that you would accidentally come upon something close enough that copyright was relevant is almost infinitesimal.
Because of this copyright is an incredibly weak restriction, and that it is as weak as it is shows clearly that any use of a copyrighted work is due to the convenience that it is available.
That is, it's about making use of the work somebody else has done, not about that restricting you somehow.
Therefore copyright is much more legitimate than ordinary property. Ordinary property, especially ownership of land, can actually limit other people. But since copyright is so sparse infringing on it is like going to world with near-infinite space and picking the precise place where somebody has planted a field and deciding to harvest from that particular field.
Consequently I think copyright infringement might actually be worse than stealing.
you're saying copying a book is worse than robbing a farmer of his food and/or livelihood, which cannot be replaced to duplicated. Meanwhile, someone who copies a book does not deprive the author of selling the book again (or a tasty proceedings from harvest).
I can't say I agree, for obvious reasons.
Just as the farmer obtains his livelihood from the investment-of-energy-to-raise-crops-to-energy cycle the author has his livelihood by the investment-of-energy-to-finding-a-useful-work-to-energy cycle.
So he is in fact robbed in a very similar way.
You'd have to steal the author's ownership of the intellectual property in order for the comparison to be valid, just as you stole ownership of his crop.
Separately, there is a reason why theft and copyright infringement are two distinct concepts in law.
Big if. Practically, the movie studios aren’t poor because their product has instances of infringement.
So the median person who is harmed when something competes with his book authorship is someone making 20k p.a., not someone who is a major shareholder of a big firm.
I remember when piracy wasn't theft, and information wanted to be free.
Ordinary property is much worse than copyright, which is both time limited and not necessarily obtained through work, and which is much more limited in availability than the number of sequences.
When someone owns land, that's actually a place you stumble upon and can't enter, whereas you're not going to ever stumble upon the story of even 'Nasse hittar en stol' (swedish 'Nasse finds a chair') a very short book for very small children.
Also, there are various incentives for teachers to publish books. Money is just one of them (I wonder how much revenue books bring to the teachers). Prestige and academic recognition is another. There are probably others still. How realistic is the depiction of a deprived teacher whose livelihood depended on the books he published once every several years?
Sayi "they have the money" is not an argument. It's about the amount of effort that is needed to individually buy, scan, process millions of pages. If that's done for you, why re-do it all?
I'm against Anthropic stealing teacher's work and discouraging them from ever writing again. Some teachers are already saying this (though probably not in California).
I think this is a fantasy. My father cowrote a Springer book about physics. For the effort, he got like $400 and 6 author copies.
Now, you might say he got a bad deal (or the book was bad), but I don't think hundreds of thousands of authors do significantly better. The reality is, people overwhelmingly write because they want to, not because of money.
Writing books is a profession.
Some people write full-time and make a living from it, through book sales, speaking gigs, teaching, or other related work.
Maybe ask Tim O’Reilly what he thinks about this so-called fantasy.
Like I said, Anthropic needs to stop stealing books or face the consequences.
My point is, the controversy is not an AI corporation vs 10^5 ordinary teachers. It's a battle of two corporations, or business models, if you will. But regardless of the result, most of the book authors will continue to get screwed, maybe the means will change. But it will not prevent them from writing, either. So I don't see any mass writers protests coming, sorry.
I also don't think Anthropic AI is going to be any less intelligent if it didn't read any modern fiction book, instead of reading a Wikipedia summary. Stories and myths are a human way of understanding the world, machines probably don't need them. And for non-fiction books - there really isn't that many irreplaceable high-profile authors out there. If it can't read, say, Feynman's Lectures on Physics, it can learn the same from 100s of other physics textbooks. Maybe they are slightly worse organized but why should superintelligence care?
Training a generative model on a book is the mechanical equivalent of having a human read the book and learn from it. Is it stealing if a person reads the book and learns from it?
Downloading the book without paying for it, which is more or less what the judge said.
Photocopying books in their entirety for commercial use is absolutely illegal.
I do not know what a person is, but I've met many people dimwitted enough that they might just be a meat LLM. LLMs aren't artificial intelligence (or even anything close), but they definitely aren't "copy machines with a blender inside"... and suggesting that they are marks you one of the dimwits I just mentioned.
Copyright maximalists are enemies of mankind.
>Photocopying books in their entirety for commercial use is absolutely illegal.
Non sequitur. This isn't "photocopying books" or anything even slightly similar. Your arguments are disingenuous.
There is no equality, and seemingly there are worker bees who can be exploited, and there are privileged ones, and of course there are the queens.
And it just so happens that that belief says they can burn whatever they want down because something in the future might happen that absolves them of those crimes.
Note: My definition of singularity isn't the one they use in San Francisco. It's the moment founders who stole the life's work of thousands of teachers finally go to prison, and their datacentres get seized.
Luigi was peanuts in comparison.
“THERE were two “Reigns of Terror,” if we would but remember it and consider it; the one wrought murder in hot passion, the other in heartless cold blood; the one lasted mere months, the other had lasted a thousand years; the one inflicted death upon ten thousand persons, the other upon a hundred millions; but our shudders are all for the “horrors” of the minor Terror, the momentary Terror, so to speak; whereas, what is the horror of swift death by the axe, compared with lifelong death from hunger, cold, insult, cruelty, and heart-break? What is swift death by lightning compared with death by slow fire at the stake? A city cemetery could contain the coffins filled by that brief Terror which we have all been so diligently taught to shiver at and mourn over; but all France could hardly contain the coffins filled by that older and real Terror—that unspeakably bitter and awful Terror which none of us has been taught to see in its vastness or pity as it deserves.”
- Mark Twain
When it comes to a lot of these teachers, I'll say, copyright work hand in hand with college and school course book mandates. I've seen plenty of teachers making crazy money off students' backs due to these mandates.
A lot of the content taught in undergrad and school hasn't changed in decades or even centuries. I think we have all the books we'll ever need in certain subjects already, but copyright keeps enriching people who write new versions of these.
Writers that have an authentic human voice and help people think about things in a new way will be fine for a while yet.
105B+ is more than Anthropic is worth on paper.
Of course they’re not going to be charged to the fullest extent of the law, they’re not a teenager running Napster in the early 2000s.
No, its not.
It's the maximum statutory damages for willful infringement, which this has not be adjudicated to be. it is not a fine, its an alternative to basis of recovery to actual damages + infringers profits attributable to the infringement.
Of course, there's also a very wide range of statutory damages, the minimum (if it is not "innocent" infringement) is $750/work.
> 105B+ is more than Anthropic is worth on paper.
The actual amount of 7 million works times $150,000/work is $1.05 trillion, not $105 billion.
Yeah, you’re probably right, I’m not a lawyer. The point is that it doesn’t matter what number the law says they should pay, Anthropic can afford real lawyers and will therefore only pay a pittance, if anything.
I’m old enough to remember what the feds did to Aaron Schwarz, and I don’t see what Anthropic did that was so different, ethically speaking.
The trial is scheduled for December 2025. That's when a jury will decide how much Anthropic owes for copying and storing those pirated books
Turns out this doesn't quite mitigate downloading them first. (Though frankly, I'm very much against people having to buy 7 million books when someone has already scanned them)
This is what every company using media are doing (think Spotify, Netflix, but also journal, ad agency, ...). I don't know why people in HN are giving a pass to AI company for this kind of behavior.
As mentioned in The Fucking Article, there's a legal difference between training an AI which largely doesn't repeat things verbatim (ala Anthropic) and redistributing media as a whole (ala Spotify, Netflix, journal, ad agency).
Because for example if you buy a movie on disc, that's a personal license and you can watch it yourself at home. But you can't like play it at a large public venue that sell tickets to watch it. You need a different and more expensive license to make money off the usage of the content in a larger capacity like that.
$500,000 per infringement...
Anthropic faces billions of dollars in damages for pirating over 7 million books to build a digital library.
The trial is scheduled for December 2025. That’s when a jury will decide how much Anthropic owes for copying and storing over 7 million pirated books
Did you read the article? The judge literally just legally recognized it.
Individuals would have their lives ruined either from massive fines or jail time.
We've held China accountable for counterfeiting products for decades and regulated their exports. So why should Anthropic be allowed to export their products and services after engaging in the same illegal activity?
We have? Are we from different multi-verses?
The one I've lived in to date has not done anything against Chinese counterfeits beyond occasionally seizing counterfeit goods during import. But that's merely occasionally enforcing local counterfeit law, a far cry from punishing the entity producing it.
As a matter of fact, the companies started outsourcing everything to China, making further IP theft and quasi-copies even easier
It's quite the mafia operation over at Amazon.
Whether or not the countermeasures have been effective in practice is a minor detail in the GP point that we would not expect an American company headquartered in the US and conducting significant business in the US to get away with the same thing.
You never noticed the hypocrite behavior all over society?
* O, you drunk drive, big fine, lots of trouble. * O, you drunk drive and are a senator, cop, mayor, ... Well, lets look the other way.
* You have anger management issues and slam somebody to the ground. Jail time. * You as a cop have anger management issues and slams somebody to the ground. Well, paid time off while we investigate and maybe a reprimand. Qualified immunity boy!
* You tax fraud for 10k, felony record, maybe jail time. * You as a exec of a company do tax fraud for 100 million. After 10 years lawyering around, maybe you get something, maybe, ... o, here is a fine of 5 million.
I am sorry but the idea of everybody being equal under the law has always been a illusion.
We are holding China accountable for counterfeiting products because it hurts OUR companies, and their income. But when its "us vs us", well, then it becomes a bit more messy and in general, those with the biggest backing (as in $$$, economic value, and lawyers), tends to win.
Wait, if somebody steal my book, i can sue that person in court, and get a payout (lawyers will cost me more but that is not the point). If some AI company steals my book, well, the chance you win is close to 1%, simply because lots of well paid lawyers will make your winning hard to impossible.
Our society has always been based upon power, wealth and influence. The more you have of it, the more you get away (or reduced) with things, that gets other fined or jailed.
Comment is more about the pseudo ethical high ground
> I honestly feel bad for some of these AI companies because the rules around copyright are changing just to target them
The ruling would be a huge win for AI companies if held. It's really weird that you reached the opposite conclusion.
Now, in theory, you learning from an author's works and competing with them in the same market could meaningfully deprive them of income, but it's a very difficult argument to prove.
On the other hand, with AI companies it's an easier argument to make. If Anthropic trained on all of your books (which is somewhat likely if you're a fairly popular author) and you saw a substantial loss of income after the release of one of their better models (presumably because people are just using the LLM to write their own stories rather than buy your stuff), then it's a little bit easier to connect the dots. A company used your works to build a machine that competes with you, which arguably violates the fair use principle.
Gets to the very principle of copyright, which is that you shouldn't have to compete against "yourself" because someone copied you.
This is one of those mental gymnastics exercises that makes copyright law so obtuse and effectively unenforceable.
As an alternative, imagine a scriptwriter buys a textbook on orbital mechanics, while writing Gravity (2013). A large number of people watch the finished film, and learn something about orbital mechanics, therefore not needing the textbook anymore, causing a loss of revenue for the textbook author. Should the author be entitled to a percentage of Gravity's profit?
We'd be better off abolishing everything related to copyright and IP law alltogether. These laws might've made sense back in the days of the printing press but they're just nonsensical nowadays.
I think, in your example, the obvious answer is no, they're not entitled to any profits of Gravity. How could you possibly prove Gravity has anything to do with someone reading, or not reading, a textbook? You can't.
However, AI participates in the exact same markets it trains from. That's obviously very different. It is INTENDED to DIRECTLY replace the things it trains on.
Meaning, not only does an LLM output directly replace the textbook it was trained on, but that behavior is the sole commercial goal of the company. That's why they're doing it, and that's the only reason they're doing it.
Maybe this is where I'm having trouble. You say "exact same markets" -- how is a print book the exact same market as a web/mobile text-generating human-emulating chat companion? If that holds, why can't I say a textbook is the exact same market as a film?
I could see the argument if someone published a product that was fine-tuned on a specific book, and marketed as "use this AI instead of buying this book!", but that's not the case with any of the current services on the market.
I'm not trying to be combative, just trying to understand.. they seem like very different markets to me.
Because the medium is actually the same. The content of a book is not paper, or a cover. It's text, and specifically the information in that text.
LLMs are intended to directly compete with and outright replace that usecase. I don't need a textbook on, say, Anatomy, because ChatGPT can structure and tell me about Anatomy, and in fact with say the exact same content slightly re-arranged.
This doesn't really hold for fictional books, nor does it hold for movies.
Watching a movie and reading a book are inherently different experiences, which cannot replace one another. Reading a textbook and asking ChatGPT about topic X is, for all intents and purposes, the same experience. Especially since, remember, most textbooks are online today.
And, I think, elephant in the room with these discussions: we cannot just compare ChatGPT to a human. That's not a foregone conclusions and, IMO, no, you can't just do that. You have to justify it.
Humans are special. Why? Because we are Human. Humans have different and additional rights which machines, and programs, do not have. If we want to extend our rights to machines, we can do that... but not for free. Oh no, you must justify that, and it's quiet hard. Especially when said machines appear to work against Humans.
So here's the thing, I don't think a textbook author going against a purveyor of online courseware has much of a chance, nor do I think it should have much of a chance, because it probably lacks meaningful proof that their works made a contribution to the creation of the courseware. Would I feel differently if the textbook author could prove in court that a substantial amount of their material contributed to the creation of the courseware, and when I say "prove" I mean they had receipts to prove it? I think that's where things get murky. If you can actually prove that your works made a meaningful contribution to the thing that you're competing against, then maybe you have a point. The tricky part is defining meaningful. An individual author doesn't make a meaningful contribution to the training of an LLM, but a large number of popular and/or prolific numbers can.
You bring up a good point, interpretation of fair use is difficult, but at the end of the day I really don't think we should abolish copyright and IP altogether. I think it's a good thing that creative professionals have some security in knowing that they have legal protections against having to "compete against themselves"
That's a point I normally use to argue against authors being entitled to royalties on LLM outputs. An individual author's marginal contribution to an LLM is essentially nil, and could be removed from the training set with no meaningful impact on the model. It's only the accumulation of a very large amount of works that turns into a capable LLM.
Paying everyone a flat rate per query is probably the only way you could do it; any other approach is either going to be contested as unfair in some way, or will be too costly to implement. But then, a flat rate is only fair if it covers everyone in proportion to the contribution, which will get diluted by the portion of training data that's not obviously attributable, like Internet comments or Wikipedia or public domain stuff or internally generated data, so I doubt authors would see any meaningful royalties from this anyway. The only thing it would do, is to make LLMs much more expensive for the society to use.
This argument would make sense if it was across the board, but it's impossible (and pretty ridiculous) to enforce in basically anything except very narrow types of media.
Let's say I come up with a groundbreaking workout routine. Some guy in the gym watches me for a while, adopts it, then goes on to become some sort of bodybuilding champion. I wouldn't be entitled to a portion of his winnings, that would be ridiculous.
Let's say I come up with a cool new fashion style. Someone sees my posts on insta and starts dressing similarly, then ends up with a massive following and starts making money in a modelling career. I wouldn't be entitled to a portion of their income, that would be ridiculous.
And yet, for some reason, media is special.
So what is the right interpretation of the law with regards to how AI is using it? What better incentivizes innovation? Do we let AI companies scan everything because AI is innovative? Or do we think letting AI vacuum up creative works to then stochastically regurgitate tiny (or not so tiny) slices of them at a time will hurt innovation elsewhere?
But obviously the real answer here is money. Copyright is powerful because monied interests want it to be. Now that copyright stands in the way of monied interests for perhaps the first time, we will see how dedicated we actually were to whatever justifications we've been seeing for DRM and copyright for the last several decades.
I really think we need to understand this as a society and also realize that moneyed interests will downplay this as much as possible. A lot of the problems we're having today are due to insufficient regulation differentiating between individuals and systems at scale.
What you're proposing is considering LLMs to be equal to humans when considering how original works are created. You could make the argument that LLM training data is no different from a human "training" themself over a lifetime of consuming content, but that's a philosophical argument that is at odds with our current legal understanding of copyright law.
> Alsup ruled that Anthropic's use of copyrighted books to train its AI models was "exceedingly transformative" and qualified as fair use, a legal doctrine that allows certain uses of copyrighted works without the copyright owner's permission.
Or even, is an individual operating within the law as fair use, the same as a voracious all-consuming AI training bot consuming everything the same in spirit?
Consider a single person in a National Park, allowed to pick and eat berries, compared to bringing a combine harvester to take it all.
Which of the following are true?
(a) the legal industry is susceptible to influence and corruption
(b) engineers don't understand how to legally interpret legal text
(c) AI tech is new, and judges aren't technically qualified to decide these scenarios
Most likely option is C, as we've seen this pattern many times before.
Any reasonable reading of the current state of fair use doctrine makes it obvious that the process between Harry Potter and the Sorcerer's Stone and "A computer program that outputs responses to user prompts about a variety of topics" is wildly transformative, and thus the usage of the copyrighted material is probably covered by fair use.
Where are you getting your data from? My conclusions are the exact opposite.
(Also, aren't judges by definition the only ones qualified to declare if it is actually fair use? You could make a case that it shouldn't be fair use, but that's different from it being not fair use.)
I think the overly liberal, non-tech crowd has become really vocal on HN as of late and your sample is likely biased by these people.
Just asking for a friend who's into this sort of thing.
Someone correct me if I am wrong but aren't these works being digitized and transformed in a way to make a profit off of the information that is included in these works?
It would be one thing for an individual to make person use of one or more books, but you got to have some special blindness not to see that a for-profit company's use of this information to improve a for-profit model is clearly going against what copyright stands for.
Although, there’s an exception for fictional characters:
https://en.m.wikipedia.org/wiki/Copyright_protection_for_fic...
Copyright isnt a digital moat. Its largely an agreement that the work is available to the public, but the creator has a limited amount of time to exploit it at market.
If you sell an AI model, or access to an AI model, theres usually around 0% of the training data redistributed with the model. You cant decompile it and find the book. As you aren't redistributing the original work copyright is barely relevant.
Imagine suggesting that because you own the design of a hammer, that all works created with the hammer belong to you and cant be sold?
That someone came up with a new method of using books as a tool to create a different work, does not entitle the original book author to a cut of the pie.
Isn't that what a lot of companies are doing, just through employees? I read a lot of books, and took a lot of courses, and now a company is profiting off that information.
Simply, if the models can think then it is no different than a person reading many books and building something new from their learnings. Digitization is just memory. If the models cannot think then it is meaningless digital regurgitation and plagiarism, not to mention breach of copyright.
The quotes "consistent with copyright's purpose in enabling creativity and fostering scientific progress." and "Like any reader aspiring to be a writer" say, from what I can tell, that the judge has legally ruled the model can think as a human does, and therefore has the legal protections afforded to "creatives."
No, that's fallacious. Using anthropomorphic words to describe a machine does not give it the same kinds of rights and affordances we give real people.
First, Authors argue that using works to train Claude’s underlying LLMs was like using
works to train any person to read and write, so Authors should be able to exclude Anthropic
from this use (Opp. 16). But Authors cannot rightly exclude anyone from using their works for
training or learning as such. Everyone reads texts, too, then writes new texts. They may need
to pay for getting their hands on a text in the first instance. But to make anyone pay
specifically for the use of a book each time they read it, each time they recall it from memory,
each time they later draw upon it when writing new things in new ways would be unthinkable.
For centuries, we have read and re-read books. We have admired, memorized, and internalized
their sweeping themes, their substantive points, and their stylistic solutions to recurring writing
problems.
...
In short, the purpose and character of using copyrighted works to train LLMs to generate
new text was quintessentially transformative. Like any reader aspiring to be a writer,
Anthropic’s LLMs trained upon works not to race ahead and replicate or supplant them — but
to turn a hard corner and create something different. If this training process reasonably
required making copies within the LLM or otherwise, those copies were engaged in a
transformative use.
[1] https://authorsguild.org/app/uploads/2025/06/gov.uscourts.ca...If a human uses a voting machine, they still have a right to vote.
Machines don't have rights. The human using the machine does.
Its like, taking notes, or google image search caching thumbnails. Honestly we dont even need the learning metaphor to see this is obviously not an infringement.
No, you can't. The only thing you can learn with is your own mind (e.g. loading notes into your laptop is not you learning).
>loading notes into your laptop is not you learning
I dont want to get too distracted by this, but tptb really hate me (questioning american notions of excellence draws the eye of sauron) and have limited my posting. Note taking is actually crucial to my actually learning things. I am largely a kinaesthetic learner, but when it comes to pure data retention, if I am not writing it out, it goes straight through. Note taking is crucial to my learning new things and retaining data, and I know I am not the only one.
Heck its a common (or was common) for writers to completely rewrite, by hand the books of "great" authors to try and learn their "voice".
Judges consider a four factor when examining fair use[1]. For search engines,
1) The use is transformative, as a tool to find content is very different purpose than the content itself.
2) Nature of the original work runs the full gamut, so search engines don't get points for only consuming factual data, but it was all publicly viewable by anyone as opposed to books which require payment.
3) The search engine store significant portions of the work in the index, but it only redistributes small portions.
4) Search engines, as original devised, don't compete with the original, in fact they can improve potential market of the original by helping more people find them. This has changed over time though, and search engines are increasingly competing with the content they index, and intentionally trying to show the information that people want on the search page itself.
So traditional search which was transformative, only republished small amounts of the originals, and didn't compete with the originals fell firmly on the side of fair use.
Google News and Books on the other hand weren't so clear cut, as they were showing larger portions of the works and were competing with the originals. They had to make changes to those products as a result of lawsuits.
So now lets look at LLMs:
1) LLM are absolutely transformative. Generating new text at users request is a very different purpose and character from the original works.
2) Again runs the full gamut (setting aside the clear copyright infringement downloading of illegally distributed books which is a separate issue)
3) For training purposes, LLMs don't typically preserve entire works, so the model is in a better place legally than a search index, which has precedent that storing entire works privately can be fair use depending on the other factors. For inference, even though they are less likely to reproduce the originals in their outputs than search engines, there are failure cases where an LLM over-trained on a work, and a significant amount the original can be reproduced.
4) LLMs have tons of uses some of which complement the original works and some of which compete directly with them. Because of this, it is likely that whether LLMs are fair use will depend on how they are being used - eg ignore the LLM altogether and consider solely the output and whether it would be infringing if a human created it.
This case was solely about whether training on books is fair use, and did not consider any uses of the LLM. Because LLMs are a very transformative use, and because they don't store original verbatim, it weighs strongly as being fair use.
I think the real problems that LLMs face will be in factors 3 and 4, which is very much context specific. The judge himself said that the plaintiffs are free to file additional lawsuits if they believe the LLM outputs duplicate the original works.
[1] https://fairuse.stanford.edu/overview/fair-use/four-factors/
Learning from the book is, well, learning from the book. Yes, they intended to make money off of that learning... but then I guess a medical student reading medical textbooks intends to profit off of what they learn from them. Guess that's not fair use either (well, it's really just use, as in the intended use for all books since they were first invented).
Once a person has to believe that copyright has any moral weight at all, I guess all rational though becomes impossible for them. Somehow, they're not capable of entertaining the idea that copyright policy was only ever supposed to be this pragmatic thing to incentivize creative works... and that whatever little value it has disappears entirely once the policy is twisted to consolidate control.
Please, please differentiate between pirating books (which Anthrophic is liable for, and is still illegal) and training on copyrighted material (which was found to be legal, for both corporations and average people).
The number of these artists I have seen receiving some bogus DMCA takedown notice for fan art is crazy.
I saw a bloke give away some of his STL's because he received a takedown request from games workshop and didnt have the funds to fight it.
Its not that I want small artists to lose, its that I want them to gain access to every bloody copyright and trademark so they are more free to create.
Shit Conde Nast managed to pull something like 400 pulps off the market, so they didnt interfere with their newly launched James Patterson collaborations.
There are many small artists who do this not for money, but for fun and have their renowned styles. Even their styles are ripped off by these generative AI companies and turned into a slot machine to earn money for themselves. These artists didn't consent to that, and this affects their (mental) well-beings.
With that context in mind, what do you think about these people who are not in this for money is ripped out of their years of achievement and their hard work exploited for money by generative AI companies?
It's not about IP (with whatever expansion you prefer) or laws, but ethics in general.
Substitute comics for any medium. Code, music, painting, illustration, literature, short movies, etc.
Yes, style copying is generally considered legal, but as another commenter posted in a related thread "scale matters".
Maybe this will be reconsidered in the near future as the scale is in a much more different level with Generative AI. While there can be no technological solution to this (since it's a social problem to begin with), maybe public opinion about this issue will evolve over time.
To be crystal clear: I'm not against the tech. I'm against abusing and exploiting people for solely monetary profit.
(2) Once you make something publicly available, anyone can learn from it. No consent necessary.
(3) Being upset does not grant you special privileges under the law.
(4) If you don't like the idea of paying for AI art, free software is both plentiful and competitive with just about anything proprietary.
(2) Learning and copying are different. If we're going to enter to the "but AI learns just like humans" realm, let's please don't, because ML tries to mimic human learning, they are not the same, even remotely.
(3) I don't think I said laws behave differently because "some humans are upset".
(4) I haven't told anything about paying for (AI) art. I don't pay for AI art, but I pay for "real" art, made by humans. I also make art sometimes, and don't use AI/ML while doing that, and don't feed models with my art (I stopped sharing on multiple platforms because of that).
(5) If we presume that every law is "correct, right and forever", then we have another name for it. It's called dogma. Moreover, we have other "unwritten" rules which generally support or sometimes supersede the laws in place, and the word for that it "ethics".
(6) Staying in the subject of law, as somebody else noted [0], scale matters. While doing style transfer for a couple of things might be considered legal, that needs a reconsideration when scale reaches to today's levels possible with generative AI.
(2) "Copying", as in verbatim using/memorizing sizeable chunks of content and suchlike? No, the goal of ML training is to generalize across concepts. I didn't make an argument based on some direct connection to neuroscience, but I also reject the idea that there isn't one.
(3) So it's an irrelevant point. Some people are upset about technology? I honestly don't care.
(4) You said that the big bad corporations are making money. I'm saying that FOSS alternatives are all over the place.
(5) That so? My code of ethics says that Luddites getting steamrolled is an immense moral good.
(6) See (1).
> My code of ethics says that Luddites getting steamrolled is an immense moral good.
Looks like it's not possible for us to discuss this further. Not because of your views, but how rigid you are on your beliefs and perspective.
With no hard feelings, I wish that time shows that you're right, all the time.
Have a nice day. :)
I do support intellectual property reform that would be considered radical by some, as I imagine you do. But my highest hopes for this situation are more modest: if AI companies are told that their data must be in the public domain to train against, we will finally have a powerful faction among capitalists with a strong incentive to push back against the copyright monopolists when it comes to the continuous renewal of copyright terms.
If the "path of least resistance" for companies like Google, Microsoft, and Meta becomes enlarging the public domain, we might finally begin to address the stagnation of the public domain, and that could be a good thing.
But I think even such a modest hope as that one is unlikely to be realized. :-\
My response to this whole thread is just “good”
Aaron Swartz is a saint and a martyr.
"Anthropic cut up millions of used books to train Claude — and downloaded over 7 million pirated ones too, a judge said."
A not-so-subtle difference.
That said, in a sane world, they shouldn't have needed to cut up all those used books yet again when there's obviously already an existing file that does all the work.
Or is it perhaps not an universal cultural/moral aspect ?
I guess for example in Europe people could be more sensitive to it.
Saying that though there are tools for digitizing books that don't require destroying them
Man of Two Worlds by Brian Herbert.
...and I did the world a favor.
https://ia800101.us.archive.org/15/items/gov.uscourts.cand.4...
Also please don't use word "learning", use "creating software using copyrighted materials".
Also let's think together how can we prevent AI companies from using our work using technical measures if the law doesn't work?
People already use pirated software for product creation.
Hypothetical:
I know a guy who learned photoshop on a pirated copy of Photoshop. He went on to be a graphic designer. All his earnings are ‘proceeds from crime’
He never used the pirated software to produce content.
If you watch a YouTube video to learn something and it's later taken down for using copyrighted images, you learned from illegal content.
I could, right now in just a few minutes, go download a perfectly functional pirated copy of nearly any Adobe program, nearly any Microsoft program and a whole range of books and movies, yet I see zero real financial troubles affecting any of the companies behind these. All the contrary in fact.
You are allowed to buy and scan books, and then used those scanned books to create products. I guess you are also allowed to pirate books and use the knowledge to create products if you are willing to pay the damages to the rights holders for copyright violations.
The whole point of copyright is to ensure you're paid for your work. AI companies shouldn't pirate, but if they pay for your work, they should be able to use it however they please, including training an LLM on it.
If that LLM reproduces your work, then the AI company is violating copyright, but if the LLM doesn't reproduce your work, then you have not been harmed. Trying to claim harm when you haven't been due to some philosophical difference in opinion with the AI company is an abuse of the courts.
People don't view moral issues in the abstract.
A better perspective on this is the fact that human individuals have created works which megacorps are training on for free or for the price of a single book and creating models which replace individuals.
The megacorps are only partially replacing individuals now, but when the models get good enough they could replace humans entirely.
When such a future happens will you still be siding with them or with individual creators?
Those damn kind readers and libraries. Giving their single copy away when they just paid for the single.
No. The point of copyright is that the author gets to decide under what terms their works are copied. That's the essence of copyright. In many cases, authors will happily sell you a copy of their work, but they're under no obligation to do so. They can claim a copyright and then never release their work to the general public. That's perfectly within their rights, and they can sue to stop anybody from distributing copies.
If the author didn't want their work to be included in an LLM, they should not have sold it, just like if an author didn't want their work to inspire someone else's work, they should not have sold it.
If that were the case then this court case would not be ongoing
I could agree with exceptions for non-commercial activity like scientific research, but AI companies are made for extracting profits and not for doing research.
> AI companies shouldn't pirate, but if they pay for your work, they should be able to use it however they please, including training an LLM on it.
It doesn't work this way. If you buy a movie it doesn't mean you can sell goods with movie characters.
> then you have not been harmed.
I am harmed because less people will buy the book if they can simply get an answer from LLM. Less people will hire me to write code if an LLM trained on my code can do it. Maybe instead of books we should start making applications that protect the content and do not allow copying text or making screenshots. ANd instead of open-source code we should provide binary WASM modules.
And the harm you describe is not a recognized harm. You don't own information, you own creative works in their entirety. If your work is simply a reference, then the fact being referenced isn't something you own, thus you are not harmed if that fact is shared elsewhere.
It is an abuse of the courts to attempt to prevent people who have purchased your works from using those works to train an LLM. It's morally wrong.
Sorry for the long quote, but basically this, yeah. A major point of free software is that creators should not have the power to impose arbitrary limits on the users of their works. It is unethical.
It's why the GPL allows the user to disregard any additional conditions, why it's viral, and why the FSF spends so much effort on fighting "open source but..." licenses.
When you do it for a transformative purpose (turning it into an LLM model) it's certainly fair use.
But more importantly, it's ethical to do so, as the agreement you've made with the person you've purchased the book from included permission to do exactly that.
Reasonable minds could debate the ethics of how the material was used, this ruling judged the usage was legal and fair use. The only problem is the material was in effect stolen.
If I didn’t license all the books I trained on, am I not depriving the publisher of revenue, given people will pay me for the AI instead of buying the book?
As the judge noted in this ruling, copyright isn’t intended to protect authors from competition. Copyright doesn’t protect Rowling from other authors writing YA wizard books cutting into her revenue streams. Or from TV producers making YA wizard shows that reduce the demand for books. Copyright doesn’t protect the Tolkien estate from Terry Brooks, or Tracy Hickman or Margret Weiss reducing the demand for Tolkien fantasy by supplanting it with their own fantasies.
> That doesn't seem to chime with the copyright notices I have read in books.
You shouldn't get your legal advice from someone with skin in the game.
I used to get scared by such verbiage. Courts ruled decades ago that many of those uses are actually permitted, under very common conditions (e.g. not distributing, etc). Yes, you totally can photocopy a book you own, for your own purposes.
But this analogy seems wrong. First, LLM is not a human and cannot "learn" or "train" - only human can do it. And LLM developers are not aspiring to become writers and do not learn anything, they just want to profit by making software using copyrighted material. Also people do not read millions of books to become a writer.
The analogy refers to humans using machines to do what would already be legally if they did it manually.
> And LLM developers are not aspiring to become writers and do not learn anything, they just want to profit by making software using copyrighted material.
[Citation needed], and not a legal argument.
> Also people do not read millions of books to become a writer.
But people do hear millions of words as children.
At a rate 1000 words/day it takes 3 years to hear a million words. Also "million words" is not equal to "million books". Humans are ridiculously efficient in learning compared to LLMs.
as long as you buy the book it still should be legal, that is if you actually buy the book and not a "read only" eBook
but the 7_000_000 pirated books are a huge issue, and one from which we have a lot of reason to believe isn't just specific to Anthropic
2020's: (Steals a bunch of books to profit off acquired knowledge.)
You are often allowed to nake a digital copy of a physical work you bought. There are tons of used, physical works thay would be good for training LLM's. They'd also be good for training OCR which could do many things, including improve book scanning for training.
This could be reduced to a single act of book destruction per copyrighted work or made unnecessary if copyright law allowed us to share others' works digitally with their licensed customers. Ex: people who own a physical copy or a license to one. Obviously, the implementation could get complex but we wouldn't have to destroy books very often.
From there, the cases would likely focus on whether that fits in established criteria for digitized copies, whether they're allowed in the training process itself, and the copyright status of the resulting model. Some countries allow all of that if you legally obtained the material in the first place. Also, they might factor whether it's for commercial use or not.
Fake altruistic mindset. Super sociopathic.
And "And only the owner can make more books" is trivially false
Keep crying about this nothingburger
"Anthropic had no entitlement to use pirated copies for its central library...Creating a permanent, general-purpose library was not itself a fair use excusing Anthropic's piracy." --- the ruling
If they committed piracy 7 million times and the minimum fine for each instance is $750 million then the law says that anthropic is liable for $5.25 billion. I just want it to be out there that they definitely broke the law and the penalty is a minimum $5.25 billion in fines according to the law, this way when none of this actually happens we at least can't pretend we didn't know.
Can anyone make a compelling argument that any of these AI companies have the public's best interest in mind (alignment/superalignment)?
Alsup detailed Anthropic's training process with books: The OpenAI rival
spent "many millions of dollars" buying used print books, which the
company or its vendors then stripped of their bindings, cut the pages,
and scanned into digital files.
I've noticed an increase in used book prices in the recent past and now wonder if there is an LLM effect in the market.Not only that but all of us are guilty too because I'm positive we've all clicked on search results that contained copyrighted content that was copied without permission. You may not have even known it was such.
Remember: Intent is irrelevant when it comes to copyright infringement! It's not that kind of law.
Intent can guide a judge when they determine damages but that's about it.
I can read 100 books and write a book based on the inspiration I got from the 100 books without any issue. However, if I pirate the 100 books I've still committed copyright infringement despite my new book being fully legal/fair use.
Ensure the models are open source, so everyone can use them, as everyones data is in there?
Close those companies and force them to delete the models, as they used copyright material?
Right guys we don't have rules for thee but not for me in the land of the free?
Some previous discussions:
https://news.ycombinator.com/item?id=44367850
Or we could recognize that’s silly when we’re talking about a group of people acting in concert and treat them as a single entity for the purpose of alleged crimes. Which is what we do when we treat a corporation as an individual for legal purposes.
This of course cannot be allowed to happen, so the the legal system is just a limbo, a bar which regular individuals must strain to pass under but that corporations regularly overstep.
As a researcher I've been furious that we publish papers where the research data is unknown. To add insult to injury, we have the audacity to start making claims about "zero-shot", "low-shot", "OOD", and other such things. It is utterly laughable. These would be tough claims to make *even if we knew the data*, simply because of its size. But not knowing the data, it is outlandish. Especially because the presumptions are "everything on the internet." It would be like training on all of GitHub and then writing your own simple programming questions to test an LLM[0]. Analyzing that amount of data is just intractable, and we currently do not have the mathematical tools to do so. But this is a much harder problem to crack when we're just conjecturing and ultimately this makes interoperability more difficult.
On top of all of that, we've been playing this weird legal game. Where it seems that every company has had to cheat. I can understand how smaller companies turn to torrenting to compete, but when it is big names like Meta, Google, Nvidia, OpenAI (Microsoft), etc it is just wild. This isn't even following the highly controversial advice of Eric Schmidt "Steal everything, then if you get big, let the lawyers figure it out." This is just "steal everything, even if you could pay for it." We're talking about the richest companies in the entire world. Some of the, if not the, richest companies to ever exist.
Look, can't we just try to be a little ethical? There is, in fact, enough money to go around. We've seen unprecedented growth in the last few years. It was only 2018 when Apple became the first trillion dollar company, 2020 when it became the second two trillion, and 2022 when it became the first three trillion dollar company. Now we have 10 companies north of the trillion dollar mark![3] (5 above $2T and 3 above $3T) These values have exploded in the last 5 years! It feels difficult to say that we don't have enough money to do things better. To at least not completely screw over "the little guy." I am unconvinced that these companies would be hindered if they had to broker some deal for training data. Hell, they're already going to war over data access.
My point here is that these two things align. We're talking about how this technology is so dangerous (every single one of those CEOs has made that statement) and yet we can't remain remotely ethical? How can you shout "ONLY I CAN MAKE SAFE AI" while acting so unethically? There's always moral gray areas but is this really one of them? I even say this as someone who has torrented books myself![4] We are holding back the data needed to make AI safe and interpretable while handing the keys to those who actively demonstrate that they should not hold the power. I don't understand why this is even that controversial.
[0] Yes, this is a snipe at HumanEval. Yes, I will make the strong claim that the dataset was spoiled from day 1. If you doubt it, go read the paper and look at the questions (HuggingFace).
[1] https://www.theverge.com/2024/8/14/24220658/google-eric-schm...
[2] https://en.wikipedia.org/wiki/List_of_public_corporations_by...
[3] https://companiesmarketcap.com/
[4] I can agree it is wrong, but can we agree there is a big difference between a student torrenting a book and a billion/trillion dollar company torrenting millions of books? I even lean on the side of free access to information, and am a fan of Aaron Swartz and SciHub. I make all my works available on ArXiv. But we can recognize there's a big difference between a singular person doing this at a small scale and a huge multi-national conglomerate doing it at a large scale. I can't even believe we so frequently compare these actions!
> In fact this business was the ultimate in deconstruction: First one and then the other would pull books off the racks and toss them into the shredder's maw. The maintenance labels made calm phrases of the horror: The raging maw was a "NaviCloud custom debinder." The fabric tunnel that stretched out behind it was a "camera tunnel...." The shredded fragments of books and magazine flew down the tunnel like leaves in tornado, twisting and tumbling. The inside of the fabric was stitched with thousands of tiny cameras. The shreds were being photographed again and again, from every angle and orientation, till finally the torn leaves dropped into a bin just in front of Robert. Rescued data. BRRRRAP! The monster advanced another foot into the stacks, leaving another foot of empty shelves behind it.
Against companies like Elsevier locking up the worlds knowledge.
Authors are no different to scientists, many had government funding at one point, and it's the publishing companies that got most of the sales.
You can disagree and think Aaron Swartz was evil, but you can't have both.
You can take what Anthropic have show you is possible and do this yourself now.
isohunt: freedom of information
If I was China I would buy every lawyer to drown western AI companies in lawsuits, because it's an easy way to win AI race.
The difference is, Aaron Swartz wasn't planning to build massive datacenters with expensive Nvidia servers all over the world.
This was the result of a cruel and zealous overreach by the prosecutor to try to advance her political career. It should never have gone that far.
The failure of MIT to rally in support of Aaron will never be forgiven.
Lesson is simple. If you want to break a law make sure it is very profitable because then you can find investors and get away with it. If you play robin hood you will be met with a hammer.
Ie. This is not a big deal. The only difference now is ppl are rapidly frothing to be outraged by the mere sniff of new tech on the horizon. Overton window in effect.