I'm running a Giraph PageRank job.  I tried with 8GB input text data
over 10 nodes (each has 4 core,  4 disks,  and 12GB physical memory),  that
is 800MB input-data/machine.    However,  Giraph job fails because of high
GC costs and Out-of-Memory exception.
    Do I set some special things in Hadoop configurations, for example,
 maximum heap size for map task vm ?

Best regards,

Reply via email to