Saw this mentioned on Slashdot a minute ago and thought of you guys:
"Cory Doctorow tells us that '[i]n 2007, John Goerzen scraped every
gopher site he could find (gopher was a menu-driven text-only precursor
to the Web; I got my first online gig programming gopher sites). He
saved 780,000 documents, totalling 40GB. Today, most of this is offline,
so he's making the entire archive available as a .torrent file; the
compressed data is only 15GB. Wanna host the entire history of a medium?
Here's your chance!' Get yourself a piece of pre-Internet history
(torrent)."
<http://tech.slashdot.org/article.pl?sid=10/04/29/2141254>
<http://www.complete.org/~jgoerzen/Full%20Gopher%20Archive%20from%202007.torrent>
Figured someone (or several someones) around here might be interested in
this... :)
Cheers,
--
Phil.
classiccmp at philpem.me.uk
http://www.philpem.me.uk/