Hi Inci,
I am not sure what's wrong. I ran the exact same command on a freshly
checked version of Graph without any trouble. Here's my output:
hadoop jar target/giraph-0.70-jar-with-dependencies.jar
org.apache.giraph.benchmark.PageRankBenchmark -e 1 -s 3 -v -V 500 -w 5
Using org.apache.giraph.benchmark.PageRankBenchmark$PageRankVertex
11/12/01 10:58:05 WARN bsp.BspOutputFormat: checkOutputSpecs:
ImmutableOutputCommiter will not check anything
11/12/01 10:58:05 INFO mapred.JobClient: Running job: job_201112011054_0003
11/12/01 10:58:06 INFO mapred.JobClient: map 0% reduce 0%
11/12/01 10:58:23 INFO mapred.JobClient: map 16% reduce 0%
11/12/01 10:58:35 INFO mapred.JobClient: map 100% reduce 0%
11/12/01 10:58:40 INFO mapred.JobClient: Job complete: job_201112011054_0003
11/12/01 10:58:40 INFO mapred.JobClient: Counters: 31
11/12/01 10:58:40 INFO mapred.JobClient: Job Counters
11/12/01 10:58:40 INFO mapred.JobClient: SLOTS_MILLIS_MAPS=77566
11/12/01 10:58:40 INFO mapred.JobClient: Total time spent by all
reduces waiting after reserving slots (ms)=0
11/12/01 10:58:40 INFO mapred.JobClient: Total time spent by all
maps waiting after reserving slots (ms)=0
11/12/01 10:58:40 INFO mapred.JobClient: Launched map tasks=6
11/12/01 10:58:40 INFO mapred.JobClient: SLOTS_MILLIS_REDUCES=0
11/12/01 10:58:40 INFO mapred.JobClient: Giraph Timers
11/12/01 10:58:40 INFO mapred.JobClient: Total (milliseconds)=13468
11/12/01 10:58:40 INFO mapred.JobClient: Superstep 3 (milliseconds)=41
11/12/01 10:58:40 INFO mapred.JobClient: Setup (milliseconds)=11691
11/12/01 10:58:40 INFO mapred.JobClient: Shutdown (milliseconds)=73
11/12/01 10:58:40 INFO mapred.JobClient: Vertex input superstep
(milliseconds)=369
11/12/01 10:58:40 INFO mapred.JobClient: Superstep 0 (milliseconds)=674
11/12/01 10:58:40 INFO mapred.JobClient: Superstep 2 (milliseconds)=519
11/12/01 10:58:40 INFO mapred.JobClient: Superstep 1 (milliseconds)=96
11/12/01 10:58:40 INFO mapred.JobClient: Giraph Stats
11/12/01 10:58:40 INFO mapred.JobClient: Aggregate edges=500
11/12/01 10:58:40 INFO mapred.JobClient: Superstep=4
11/12/01 10:58:40 INFO mapred.JobClient: Last checkpointed superstep=2
11/12/01 10:58:40 INFO mapred.JobClient: Current workers=5
11/12/01 10:58:40 INFO mapred.JobClient: Current master task partition=0
11/12/01 10:58:40 INFO mapred.JobClient: Sent messages=0
11/12/01 10:58:40 INFO mapred.JobClient: Aggregate finished vertices=500
11/12/01 10:58:40 INFO mapred.JobClient: Aggregate vertices=500
11/12/01 10:58:40 INFO mapred.JobClient: File Output Format Counters
11/12/01 10:58:40 INFO mapred.JobClient: Bytes Written=0
11/12/01 10:58:40 INFO mapred.JobClient: FileSystemCounters
11/12/01 10:58:40 INFO mapred.JobClient: FILE_BYTES_READ=590
11/12/01 10:58:40 INFO mapred.JobClient: HDFS_BYTES_READ=264
11/12/01 10:58:40 INFO mapred.JobClient: FILE_BYTES_WRITTEN=129240
11/12/01 10:58:40 INFO mapred.JobClient: HDFS_BYTES_WRITTEN=55080
11/12/01 10:58:40 INFO mapred.JobClient: File Input Format Counters
11/12/01 10:58:40 INFO mapred.JobClient: Bytes Read=0
11/12/01 10:58:40 INFO mapred.JobClient: Map-Reduce Framework
11/12/01 10:58:40 INFO mapred.JobClient: Map input records=6
11/12/01 10:58:40 INFO mapred.JobClient: Spilled Records=0
11/12/01 10:58:40 INFO mapred.JobClient: Map output records=0
11/12/01 10:58:40 INFO mapred.JobClient: SPLIT_RAW_BYTES=264
Would it be possible to send me the logs from the first attempts for
every map task?
i.e. from
Task attempt_201111302343_0002_m_000000_0
Task attempt_201111302343_0002_m_000001_0
Task attempt_201111302343_0002_m_000002_0
Task attempt_201111302343_0002_m_000003_0
Task attempt_201111302343_0002_m_000004_0
Task attempt_201111302343_0002_m_000005_0
I think that could help us find the issue.
Thanks,
Avery
On 12/1/11 1:17 AM, Inci Cetindil wrote:
Hi,
I'm running PageRank benchmark example on a cluster with 1 master + 5 slave
nodes. I have tried it with a large number of vertices; when I failed I decided
to make it run with 500 vertices and 5 workers first. However, it doesn't work
even for 500 vertices.
I am using the latest version of Giraph from the trunk and running the
following command:
hadoop jar giraph-0.70-jar-with-dependencies.jar
org.apache.giraph.benchmark.PageRankBenchmark -e 1 -s 3 -v -V 500 -w 5
I attached the error message that I am receiving. Please let me know if I am
missing something.
Best regards,
Inci