Zhuang Kechen created HAMA-640:
----------------------------------
Summary: Large graph computing cause out of memory
Key: HAMA-640
URL: https://issues.apache.org/jira/browse/HAMA-640
Project: Hama
Issue Type: Question
Components: graph
Affects Versions: 0.5.0
Environment: hadoop-1.0.3 hama-0.5.0
Reporter: Zhuang Kechen
When I test some small graphs PageRank on distributed environment, everything
went all right. When I upload a Larger graph(758Mb) on HDFS, the graph format
is all the same, and the PageRank went right on SuperStep 0,1,2, then job
failed after that, out of memory. I didn't change the examples, the test use
the hama-examples pagerank.
Somebody please help...Thanks!!
--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira