thread convergence! It didn't include wikipedia-proper when I looked yesterday, but this was suggested...
On Tue, Feb 24, 2009 at 11:26 PM, Brian <[email protected]> wrote: > Why not make the uncompressed dump available as an Amazon Public > Dataset? http://aws.amazon.com/publicdatasets/ On Wed, Feb 25, 2009 at 10:43 AM, Nathan <[email protected]> wrote: > http://www.nytimes.com/external/readwriteweb/2009/02/25/25readwriteweb-amazon_exposes_1_terrabyte_of.html > > According to this, a new project by Amazon that makes a terabyte of public > data available includes a full dump of Wikipedia. It also includes the > complete dbpedia - so it seems like there are likely to be lots of > duplicates. Given the other information it says it includes (the whole human > genome, all other publicly available DNA sequences, census data, etc.) I'm > not sure how it all fits in a single terabyte. Interesting concept, though. > I wonder how old the dump is, since they've been unavailable for some time? > > Nathan > > -- > Your donations keep Wikipedia running! Support the Wikimedia Foundation > today: http://www.wikimediafoundation.org/wiki/Donate > _______________________________________________ > foundation-l mailing list > [email protected] > Unsubscribe: https://lists.wikimedia.org/mailman/listinfo/foundation-l > > _______________________________________________ foundation-l mailing list [email protected] Unsubscribe: https://lists.wikimedia.org/mailman/listinfo/foundation-l
