Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I haven't seen much discussion in defense of Reddit protecting their content from LLM training competitors. This to me is why they have to crack down on their API, it's no longer just SEO links back, it's training someone else's models on your content and community for free. This to me is the elephant. It's horrible how they treat their app community, but this is a massive problem for them.


That's already happened; if that were the reason, they'd be trying to close the barn after the horses have already crossed into another state.


it's not a one time scrape, but a continual tuning


Even if we ignore the idea of just scraping the site, how much would it cost an API user to grab most posts just once? Is it actually enough to stop anyone?




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: