Add below property to hama-site.xml and retry please.
<property>
<name>hama.graph.multi.step.partitioning.interval</name>
<value>3000000</value>
</property>
P.S., keep in mind that graph job runs in memory so there's a capacity limit.
On Tue, Oct 23, 2012 at 3:57 PM, Shuo Wang <[email protected]> wrote:
> Yes! I have used the data on the Stanford SNAP,the largest data is 106M,
> they all work.
>
> 2012/10/23 Edward J. Yoon <[email protected]>
>
>> Hi,
>>
>> If data is smaller than 100M, it works?
>>
>> On Tue, Oct 23, 2012 at 12:04 PM, Shuo Wang <[email protected]>
>> wrote:
>> > HI,
>> >
>> > I have done some PageRank experiments on HAMA, but when I used the data
>> > larger than 110M,it failed! Our cluster has 10 nodes, 45 tasks, each task
>> > has 1G memory.
>> > Here is the output:
>> >
>> > 12/10/22 14:56:31 INFO bsp.FileInputFormat: Total input paths to process
>> :
>> > 45
>> > 12/10/22 14:56:31 INFO bsp.BSPJobClient: Running job:
>> job_201210221428_0003
>> > 12/10/22 14:56:34 INFO bsp.BSPJobClient: Current supersteps number: 0
>> > 12/10/22 14:56:43 INFO bsp.BSPJobClient: Current supersteps number: 2
>> > 12/10/22 14:57:46 INFO bsp.BSPJobClient: Job failed.
>>
>>
>>
>> --
>> Best Regards, Edward J. Yoon
>> @eddieyoon
>>
--
Best Regards, Edward J. Yoon
@eddieyoon