Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

from the blog: >>> First of all, it compresses not the entireity, but rather the most popular subset of the English Wikipedia. Two dumps are distributed at time of writing, the top 1000 articles and the top 300,000 requiring approximately 10MB and 1GB, respectively.


Actually the top 1337 and 314159 articles, respectively :).


Cheers! Then their pitch of "nearly all content" is grossly wrong. Meh.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: