Thanks! :D

OK, I'll do as soon as possible!.

2011/3/21 Miklós Erdélyi <[email protected]>:
> Hi Edward!
>
> You convinced me -- let's go ahead with the 0.2 release without the
> patch in question ;-)
>
> But please commit the patch as soon as possible to make it easier to
> do incremental improvements upon it!
>
> Thanks,
> Miklos
>
>
> 2011/3/18 Edward J. Yoon <[email protected]>:
>> Hi Miklos!
>>
>> Yes, I've already looked at your patch. It's perfect! (actually it was
>> on my todo list.)
>>
>> By the way, having been developing the 0.2 for so long, I think it is
>> time to release the Hama 0.2 and implement next features. (You can
>> check our roadmap here http://wiki.apache.org/hama/RoadMap)
>>
>> Here's more detailed my thoughts and opinions about 0.2 release:
>>
>> 1) Currently, the Apache Hama 0.2 provides a *basic* functionality
>> enabling the user to understand the Bulk Synchronous Parallel. That's
>> it! so there's no performance improvement plan anymore. I think, ...
>> it'll also provide big participate opportunities for new volunteers
>> like you. :)
>>
>> 2) Once we release Hama 0.2, we can just start the Hama 0.3. IMO,
>> It'll be a *product level* software includes many useful features with
>> improved performance, such as In/output system, Runtime compression of
>> BSP message, Network condition based mechanisms, Web UI, ..., etc.
>>
>> As above reason, I prefer to commit your patch to 0.3 version.
>>
>> What do you think?
>>
>> P.S. I hope you forgive me, I'm CC our community to discuss this together.
>>
>> 2011/3/18 Miklós Erdélyi <[email protected]>:
>>> Hi Edward,
>>>
>>> Thanks for accepting my first little patch for HAMA (TestBSPPeer 
>>> modification)!
>>>
>>> I'm a PhD student in Hungary at the University of Pannonia. At the
>>> moment my research topic is related to link-based graph-similarity
>>> algorithms for which I plan to do measurements on implementations in
>>> HAMA BSP.
>>> I've already written a prototype on top of HAMA which supports
>>> Pregel-like graph computations without fault-tolerance. It supports
>>> loading graph partitions by peers, performing local computations on
>>> vertexes, grouping of incoming messages per vertex and stopping after
>>> a fixed number of iterations.
>>>
>>> Currently the greatest bottleneck is the messaging part: there I would
>>> like to help improving HAMA. I'd like to implement asynchronous
>>> message processing and grouping incoming messages by message tag. The
>>> latter would allow more efficient iteration through messages intended
>>> for a specific vertex while the former would shorten the computational
>>> + messaging phase generally.
>>>
>>> As a first step I've uploaded to JIRA a simple patch which speeds up
>>> HAMA messaging by delivering messages for the local peer directly into
>>> its queue (i.e., there's no network transfer for locally delivered
>>> messages). Also, this patch adds a peer InetNetworkAddress cache to
>>> avoid costly InetNetworkAddress construction on every message sent.
>>> You have already commented on this patch so I guess you have already
>>> looked at it.
>>>
>>> In case you have ideas or pointers for the HAMA-based graph
>>> computation framework in general or improving messaging performance in
>>> particular, please let me know! :)
>>>
>>> Thanks,
>>> Miklos
>>>
>>
>>
>>
>> --
>> Best Regards, Edward J. Yoon
>> http://blog.udanax.org
>> http://twitter.com/eddieyoon
>>
>



-- 
Best Regards, Edward J. Yoon
http://blog.udanax.org
http://twitter.com/eddieyoon

Reply via email to