Hmm... code: *check* desire to add stochastic decomp to code: *check* amazon credits: *check* (my account today: almost $300 left burning hole in pocket) relatively gigantic social graph: *check* legal ability to put gigantic social graph on ec2: not so check, but maybe some clever anonymization work on export could be done here.
Let's break some records! :) -jake On Thu, Feb 25, 2010 at 12:18 PM, Drew Farris <drew.far...@gmail.com> wrote: > Sound's pretty interesting. Assuming this is EC2, Would be great if > Amazon would pick up the tab, us being an open source project and all > and potentially good marketing to boot. Also, whomever's account is > used will have to have its default limit of 20 machines raised. > > On Thu, Feb 25, 2010 at 3:11 PM, Robin Anil <robin.a...@gmail.com> wrote: > > +1 I'm ready. What do we need. Perf Tuning! Cluster Setup?, Amazon > Credits? > > Someone to pay for the machines or from our own pockets? > > > > > > Robin > > > > On Fri, Feb 26, 2010 at 1:20 AM, Ted Dunning <ted.dunn...@gmail.com> > wrote: > > > >> These guys: > >> > >> > >> > http://delivery.acm.org/10.1145/1460000/1459718/a18-vigna.pdf?key1=1459718&key2=4070317621&coll=GUIDE&dl=GUIDE&CFID=77555530&CFTOKEN=13940667 > >> > >> say this: > >> > >> > We present experiments over a collection with 3.6 billions of > >> postings---two orders of magnitudes larger than any published experiment > in > >> the literature. > >> > >> My impression is that Mahout on about 100 machines is ready to break > this > >> record with Jake's latest code. The stochastic decomposition should > make > >> it > >> even more plausible. > >> > >> The hardest part will be to find reasonable data with > 4 billion > non-zero > >> entries. At 0.01% sparsity, this is roughly a square matrix with 5 > million > >> rows and columns. > >> > >> Jake, your social graph should be much larger than that. > >> > >> -- > >> Ted Dunning, CTO > >> DeepDyve > >> > > >