Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

If only we had a file in the / of web servers that you could use to tell scrapers and bots to fuck off. We'd say for instance:

     User-Agent: *
     Disallow: /
And that would be that. Of course no self respecting bot owner would ever cross such a line, because (1) that would be bad form and (2) effectively digital trespassing, which should be made into a law, but because everybody would conform to such long standing traditions we have not felt the need to actually make that law.


>which should be made into a law

1. People in other countries probably don't give a fuck about your laws, global internet and all.

2. How are you going to define this law in such a manner that isn't going to be a problem for someone, for example, writing a plugin in the browser to manipulate the page for their own personal reasons.... 'scraping' is a very broad term that can easily include viewing.


Harmonization is a thing.

We've done it for lots of other things, I don't see why it would not work for the #1 technological critical resource.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: