Thanks! :D OK, I'll do as soon as possible!.
2011/3/21 Miklós Erdélyi <[email protected]>: > Hi Edward! > > You convinced me -- let's go ahead with the 0.2 release without the > patch in question ;-) > > But please commit the patch as soon as possible to make it easier to > do incremental improvements upon it! > > Thanks, > Miklos > > > 2011/3/18 Edward J. Yoon <[email protected]>: >> Hi Miklos! >> >> Yes, I've already looked at your patch. It's perfect! (actually it was >> on my todo list.) >> >> By the way, having been developing the 0.2 for so long, I think it is >> time to release the Hama 0.2 and implement next features. (You can >> check our roadmap here http://wiki.apache.org/hama/RoadMap) >> >> Here's more detailed my thoughts and opinions about 0.2 release: >> >> 1) Currently, the Apache Hama 0.2 provides a *basic* functionality >> enabling the user to understand the Bulk Synchronous Parallel. That's >> it! so there's no performance improvement plan anymore. I think, ... >> it'll also provide big participate opportunities for new volunteers >> like you. :) >> >> 2) Once we release Hama 0.2, we can just start the Hama 0.3. IMO, >> It'll be a *product level* software includes many useful features with >> improved performance, such as In/output system, Runtime compression of >> BSP message, Network condition based mechanisms, Web UI, ..., etc. >> >> As above reason, I prefer to commit your patch to 0.3 version. >> >> What do you think? >> >> P.S. I hope you forgive me, I'm CC our community to discuss this together. >> >> 2011/3/18 Miklós Erdélyi <[email protected]>: >>> Hi Edward, >>> >>> Thanks for accepting my first little patch for HAMA (TestBSPPeer >>> modification)! >>> >>> I'm a PhD student in Hungary at the University of Pannonia. At the >>> moment my research topic is related to link-based graph-similarity >>> algorithms for which I plan to do measurements on implementations in >>> HAMA BSP. >>> I've already written a prototype on top of HAMA which supports >>> Pregel-like graph computations without fault-tolerance. It supports >>> loading graph partitions by peers, performing local computations on >>> vertexes, grouping of incoming messages per vertex and stopping after >>> a fixed number of iterations. >>> >>> Currently the greatest bottleneck is the messaging part: there I would >>> like to help improving HAMA. I'd like to implement asynchronous >>> message processing and grouping incoming messages by message tag. The >>> latter would allow more efficient iteration through messages intended >>> for a specific vertex while the former would shorten the computational >>> + messaging phase generally. >>> >>> As a first step I've uploaded to JIRA a simple patch which speeds up >>> HAMA messaging by delivering messages for the local peer directly into >>> its queue (i.e., there's no network transfer for locally delivered >>> messages). Also, this patch adds a peer InetNetworkAddress cache to >>> avoid costly InetNetworkAddress construction on every message sent. >>> You have already commented on this patch so I guess you have already >>> looked at it. >>> >>> In case you have ideas or pointers for the HAMA-based graph >>> computation framework in general or improving messaging performance in >>> particular, please let me know! :) >>> >>> Thanks, >>> Miklos >>> >> >> >> >> -- >> Best Regards, Edward J. Yoon >> http://blog.udanax.org >> http://twitter.com/eddieyoon >> > -- Best Regards, Edward J. Yoon http://blog.udanax.org http://twitter.com/eddieyoon
