That's just a win-win situation, you're using their services for free because it helps you, they use your interaction to improve the model; the model is still free to use.
There's no win win situation. My content is stolen and given to others. I've lost. Google paid me for traffic via ads, therefore I allowed google to ingest my content. You as a person could read it. I've never given you permission to resell it, and if you did, I'd come after you to pay royalties. The same must apply to openai and other leeches.
Physical property is either borrowed, owned, sold, and so on.
If your spouse takes your car to work without your knowledge it's borrowed.
If they take it and sell it without consent it's theft.
Same applies to data. But data is electrons and as such it can't be moved, it is "copied". So technically speaking you are right, but practically you are not. If you steal NBC's prerelease movie then that's theft. As is copying it without constent. Once you pay for it you can copy it from their servers to your device. But you can't copy it to someone else's machine.
> If you steal NBC's prerelease movie then that's theft.
No. Advocates of expanded IP law have attempted to spread the idea that copyright infringement is "theft" as it adds emotional weight to their arguments. "You wouldn't download a car" etc. Same for the use of the word "piracy" - borrow an emotionally laden term from another context and hope nobody notices the sleight of hand.
And it's important that we reject this definition because it distorts the reality of the situation.
> And it's important that we reject this definition because it distorts the reality of the situation.
Depends who's reality. A content creator's reality is that their content is indeed stolen and monetised by someone without permission.
"Advocates of expanded IP law" do appear to be in the right, at least by law. Copying and distributing digital products is treated more or less as theft, particularly when done at scale.
AI and current training practices are even worse than stealing someone's work. It steals someone's identity. AI can copy unique characteristics, not just individual content to reproduce identical content. It can replicate a person's unique style without consent, and that's uniquely dangerous.
On a trivial level this is correct as words mean what we collectively decide they mean.
However I am making the point that a) the meaning has been changed and b) it has changed in a way that is deceptive and masks a useful fact about the world
> On a trivial level this is correct as words mean what we collectively decide they mean.
Correct, and collectively we decided that reselling digital work without permission is indeed theft, just as we rightfully decided that digital goods for the most part are like physical goods.
> a) the meaning has been changed
It hasn't really, digital theft still has the same meaning as any form of theft. Some did try to change the meaning and non trivialise the act based on the fact that digital goods are not like phyisical goods. But that's a techincallity based on the nature of digital goods.
Similarly, AI folks wish to change the meaning of theft based on the false assumption that an AI system "learns just like a human". But that's a false assumption. The software does mimic human behaviour, but we all know that it is neither human nor intelligent (if it were intelligent you'd show it a set of multiplications, and from that point onwards it would figure it out on its own. same with writing stories). Yet some are trying to change the meaning of words to accommodate their view of the world in which software that can ingest people's IP at massive scale, mix it in, and output something that looks novel is somehow similar to human learning.
Therefore the matter is trivial. Software ingestsing digital content without permission, and outputting content made of even tiny bits of the original, is theft. Simple as that. However, that does not mean that AI should be banned. It's how the AI software is fed its data that must be brought in line.
This debate predates modern AI and I've been having this debate for a lot longer than generative AI had been around. I think it's more likely that you really want to make a point about AI rather then you have deeply held views on intellectual property
It would be a win-win if the company promised that they'll keep the AI as it is, and as free as it is, as long as the company functions. Then they would take something, give something, and we could discuss if what we get outweighs what they took.
But the street is one-way, and it's the company that has the upper hand. The company can (and does) retract access to the AI, but they themselves keep what they took. If in the meantime people became attached to what the company gave, the company even does damage to them, not just by taking away the access, but because of severing the supply for a dependency.
So the people are taken advantage of because the company took the assets, they are taken advantage of because they help to further train the AI by using it, and then they get, at most, the privilege to pay for something that grew out of them.
That's why it's not a win-win. It's a win for the company, and a questionable outcome, and a risk for the people.
I agree wrt/ experience, but I don't think it applies to this situation. Even if you had an experience that would end, their ownership of the data wouldn't, and that, among other things, make this very one-sided.
I do want to stress something from your conclusion though. That people do better if they anticipate change, and can adapt to it.
Whether it's one-sided depends on what you think you've gained and lost. I publish code for free (open source) and I publish my writing for free (on my blog and as comments on various websites).
I don't expect compensation from anyone who uses them, whether it's public or private use, so I don't feel like I've lost anything. Sometimes people "pay it forward." If I actually get something back, that's a win.
There are web search engines and AI chatbots that might be very slightly better (unmeasurably so) due to having been trained on stuff I published over the years. Meanwhile I get a lot of benefit from using free stuff on the Internet. I think that's a one-sided deal in my favor.
(I also pay for GPT4 access. Whether it's worth $20 a month is more questionable, but it's fun to play with and so far I'm interested enough that I haven't cancelled.)
>Whether it's one-sided depends on what you think you've gained and lost.
I completely agree. At the end of the day, winning and losing in this situation cannot be measured, especially the "losing" part wrt/ people, so it all boils down to how the individuals perceive it. (Which is of course why powerful entities put so much effort into PR.)
I personally feel better if there are some safeguards around usage, and so I like licenses like the GPL family, where regulations are in place so that the effort is not completely trivially closed up.
But really, at the end of the day what we can control best is our perception of thing. Life is what we make of it.