Sometimes I just feel like these people overestimate how much they are actually owed from these training runs.
It’s trained on 15T tokens. So how many did you provide that were genuinely novel? And how much money do you want? Like $5 from OpenAI? And $0 from meta since it’s open source?
I personally hope we can all get on the same team with AI and treat its advancement as scientific research for the betterment of humanity.
> It’s trained on 15T tokens. So how many did you provide that were genuinely novel?
Are we suggesting that we should ditch creators' rights and instead value intellectual property along the lines of "I should be able to copy all your stuff as long as long as I copy lots of other stuff too, and give it all away for free or almost free...?"
That sounds like a pretty good deal to me - but I've always believed that the entire concept of "intellectual property" does overall more harm than good.
> I've always believed that the entire concept of "intellectual property" does overall more harm than good
It's fairly broken, but on balance it seems the creators are the ones getting screwed.
I did years of research in a scientific lab which resulted in <drum roll> 2 (yes two ... count them) peer-reviewed papers.
My colleagues and I did the work, wrote up the damned papers, yet to get them published we had to sign over copyright to what I'd now suggest is essentially a rent-seeking scientific publishing mafia.
All a long time ago, but I never had (and still don't have) the ability to either legally download or legally redistribute my own work...
The point (often) is to stop the practice, not to ask money for it.
Like, you get fined for speeding, but if you keep speeding you'll get you're license revoked, and if you keep driving after that you get jail time. The payment required is punitive, but the point is to stop you.
Again, that an opinion that will need to be tested in the courts.
You will have to explain why Hunter Thompson copying every word of every Hemingway novel isn't copyright infringement, but a computer doing the same is.
1. Humans are not machines; arguments saying because a human can learn LLMs must be allowed to copy is not interesting.
2. Did Thompson publish the work? It sounds like you're referring to an activity Thompson did in private, to improve his skills as an author. Meanwhile, lawsuits are alleging that LLM services reproduce copyrighted materials.
3. What can be fair use at small scale is no longer fair use at large scale.
Of course. The alternative is that creators dictate the price for any of the infinite number of zero cost copies which is and always has been ridiculous.
Which they do. That’s what the New York Times lawsuit is about. And in Meta’s case, they went specifically out of their way to remove the copyright notices to hide their actions.
> I personally hope we can all get on the same team with AI and treat its advancement as scientific research for the betterment of humanity.
s/AI/capital/.
It's painfully obvious that this is going to make material conditions worse for most people who use their minds to work instead of their hands. to these people, the "betterment of humanity" is a cruel joke.
The normal way to figure this out is to negotiate. We’ll either come to a mutually agreeable amount, or they’ll decide it’s not worth the cost to use my stuff. If I think I deserve $5 from OpenAI, then I’d suggest that, and they’d accept or come back with a counteroffer or tell me I’m nuts and move on. Probably that last one.
But for some reason, these companies think they don’t need to bother, and can just use everyone’s stuff.
Wait, I phrased that wrong. For a very good reason based on long precedent, these companies know that IP law is a tool to be used by big companies against individuals and sometimes other big companies, but never by individuals against company, so they know they don’t have to bother.
> Sometimes I just feel like these people overestimate how much they are actually owed from these training runs.
It’s not about being paid for including their work, it’s about being compensated for having done so without permission. For crying out loud, they went out of their way to remove copyright notices from the pirated work.
> It’s trained on 15T tokens. So how many did you provide that were genuinely novel?
Then they can just take it out. And go ahead and take out every thing you didn’t have permission to include. What’s that? The model is now significantly worse? Yeah, these things compound.
> And how much money do you want? Like $5 from OpenAI? And $0 from meta since it’s open source?
No, they would’ve wanted for the work to not have been included without permission in the first place. Do you understand the world you’re advocating for? You’re arguing it’s OK for rich people to do whatever they want if they throw some scraps on the floor for you. Not everything is about money. Unfortunately there’s no other reasonable way (legal and non violent) to punish these infringers.
> I personally hope we can all get on the same team with AI and treat its advancement as scientific research for the betterment of humanity.
What you’re expressing is “I hope everyone will stop arguing and agree with me”. These moguls care about themselves, it is incredibly naive to believe they give a rat’s ass about “the betterment of humanity”.
LLM companies aren't being funded by the hundreds of billions because investors expect science to be advanced by text and image generators.
I find it very unlikely that the commodification of knowledge work will be for the betterment humanity, I don't know if people are expecting here that just because the value of more people's labor becomes zero that we will do, what, do away with money? No, it will just mean that fewer people will have the chance to earn the right to use space and resources in a meaningful way.
Hard to treat llms training aon your data at your expense as research for the betterment of humanity when it is specifically the private company who is imposing that cost on you that profits.
Does this goes both ways? Can I infringe of Disney’s IP on the grounds that their stories are so derivative that they aren’t actually that new?
The betterment of humanity seems to involve some parties making a ton of money while the people who provided the data apparently just need to be grateful.
The fact they they were willing to risk significant legal exposure in order to use this dataset suggests it's worth considerably more than 5 dollars to them. Zuck isn't putting his ass on the line for a Big Mac.
Assuming the works have registered with the copyright office they're eligible for statutory damages.
The range for that is huge though, it can be in the hundreds of dollars per work, or if the infringement is shown to be wilful then a judge can award up to $150,000 per work.
It’s trained on 15T tokens. So how many did you provide that were genuinely novel? And how much money do you want? Like $5 from OpenAI? And $0 from meta since it’s open source?
I personally hope we can all get on the same team with AI and treat its advancement as scientific research for the betterment of humanity.