That was of great help. Thanks for the input. There is something called*MARS
*for accelerating MapReduce using GPUs. I am quite not sure about this. But
here's the link <http://www.cse.ust.hk/gpuqp/Mars_tr.pdf>. This paper might
give you a better idea of what we are trying to achieve. And we are
thinking of using JCUDA for parallelizing things on Nividia Tesla. We are
still at Ground zero. But if we can accelerate the performance of mahout's
k-means on GPU, i guess it will be a huge thing.

P.S : Perhaps the paper would transform the artificial marriage to
something better


On Wed, Feb 20, 2013 at 4:08 PM, Sean Owen <[email protected]> wrote:

> I think this is quite possible too. I just think there's little point in
> matching this up with Hadoop. They represent entirely different
> architectures for large-scale computation. I mean, you can probably write
> an M/R job that uses GPUs on workers, but I imagine it would be an
> artificial marriage of technologies. Probably Hadoop being used simply to
> distribute data.
>
> If you want to use a GPU, and want to use it properly, most of your work is
> to create an effective in-core parallel implementation, not distributed
> across computers and distributed file systems. You use JNI or CUDA bindings
> in Java to push computations into hardware from Java.
>
> This is an exercise in a) modifying a matrix/vector library to use native
> hardware, then b) writing algorithms that use that library. I think your
> best starting point in Java may be something more general like Commons
> Math.
>
>
>
>
> On Wed, Feb 20, 2013 at 10:22 AM, 万代豊 <[email protected]> wrote:
>
> > This is the agenda that I'm interested in too.
> > I believe Item-Based Recomemndation in Mahout (Not only about Mahout
> > though) should spend sometime
> > doing multiplication of cooccurrence matrix and user preference vector.
> > If we could pass this multiplication task off loaded to GGPU, then that
> > will be a great acceleration.
> > What I'm not really clear is how double precision multiplication task
> > inside Java Virtual Machine can take advantage of the HW accelerator.(I
> > mean how can you make GGPU visible to Mahout through JVM?)
> >
> > If we could get over this in addition to what Ted Dunning presented the
> > other day on Solr involment in building/loading cooccurrence matrix for
> > Mahout recommendation, it should be a big leap in innovating Mahout
> > recommendation.
> >
> > Am I missing sothing or just dreamig?
> > Regards,,,
> > Y.Mandai
> >
> > 2013/2/20 Sean Owen <[email protected]>
> >
> > > I think all of the code uses double-precision floats. I imagine much of
> > it
> > > could work as well with single-precision floats.
> > >
> > > MapReduce and a GPU are very different things though, and I'm not sure
> > how
> > > you would use both together effectively.
> > >
> > >
> > > On Wed, Feb 20, 2013 at 7:10 AM, shruti ranade <
> [email protected]
> > > >wrote:
> > >
> > > > Hi,
> > > >
> > > > I am a beginner in mahout. I am working on k-means MR implementation
> > and
> > > > trying to run it on a GPGPU.* I wanted to know if mahout computations
> > are
> > > > all double precision or single precision. *
> > > >
> > > > Suggest me any documentation that I need to refer to.
> > > >
> > > > Thanks,
> > > > Shruti
> > > >
> > >
> >
>

Reply via email to