Hacker News new | past | comments | ask | show | jobs | submit login

>"After reading this article, you’ll be able to save all ~6 million pages of Wikipedia so you can access the sum of human knowledge regardless of internet connection!"

[...]

>"The current Wikipedia file dump in English is around 95 GB in size. This means you’ll need something like a 128 GB flash drive to accommodate the large file size."

Great article!

Also, on a related note, there's an interesting philosophical question related to this:

Given the task of preserving the most important human knowledge from the Internet and given a certain limited amount of computer storage -- what specific content (which could include text, pictures, web pages, PDFs, videos, technical drawings, etc.) from what sources do you select, and why?

?

So first with 100GB (All of Wikipedia is a great choice, btw!) -- but then with only 10GB, then 1GB, then 100MB, then 10MB, then 1MB, etc. -- all the way down to 64K! (about what an early microcomputer could hold on a floppy disk...)

What information do you select for each storage amount, and why?

?

(Perhaps I should make this a future interview question at my future company!)

Anyway, great article!




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: