Oh, so it is the task running on the other node (ncdm-15) fails and Hadoop re-run the task on the local node (ncdm-8). (I only have two nodes, ncdm-8 and ncdm-15. Both namenode and jobtracker are running on ncdm-8. The program is also started on ncdm-8).

08/01/18 19:08:27 INFO mapred.JobClient: Task Id : 
task_200801181852_0001_m_000001_0, Status : FAILED Too many fetch-failures
08/01/18 19:08:27 WARN mapred.JobClient: Error reading task outputncdm15

Any ideas why the task would fail? And why it takes so long for Hadoop to detect the failure?

Thanks
Yunhong

On Sat, 19 Jan 2008, Devaraj Das wrote:

Hi Yunhong,
As per the output it seems the job ran to successful completion (albeit with
some failures)...
Devaraj

-----Original Message-----
From: Yunhong Gu1 [mailto:[EMAIL PROTECTED]
Sent: Saturday, January 19, 2008 8:56 AM
To: hadoop-user@lucene.apache.org
Subject: Re: Reduce hangs



Yes, it looks like HADOOP-1374

The program actually failed after a while:


[EMAIL PROTECTED]:~/hadoop-0.15.2$ ./bin/hadoop jar
hadoop-0.15.2-test.jar mrbench
MRBenchmark.0.0.2
08/01/18 18:53:08 INFO mapred.MRBench: creating control file:
1 numLines, ASCENDING sortOrder
08/01/18 18:53:08 INFO mapred.MRBench: created control file:
/benchmarks/MRBench/mr_input/input_-450753747.txt
08/01/18 18:53:09 INFO mapred.MRBench: Running job 0:
input=/benchmarks/MRBench/mr_input
output=/benchmarks/MRBench/mr_output/output_1843693325
08/01/18 18:53:09 INFO mapred.FileInputFormat: Total input
paths to process : 1
08/01/18 18:53:09 INFO mapred.JobClient: Running job:
job_200801181852_0001
08/01/18 18:53:10 INFO mapred.JobClient:  map 0% reduce 0%
08/01/18 18:53:17 INFO mapred.JobClient:  map 100% reduce 0%
08/01/18 18:53:25 INFO mapred.JobClient:  map 100% reduce 16%
08/01/18 19:08:27 INFO mapred.JobClient: Task Id :
task_200801181852_0001_m_000001_0, Status : FAILED Too many
fetch-failures
08/01/18 19:08:27 WARN mapred.JobClient: Error reading task
outputncdm15
08/01/18 19:08:27 WARN mapred.JobClient: Error reading task
outputncdm15
08/01/18 19:08:34 INFO mapred.JobClient:  map 100% reduce 100%
08/01/18 19:08:35 INFO mapred.JobClient: Job complete:
job_200801181852_0001
08/01/18 19:08:35 INFO mapred.JobClient: Counters: 10
08/01/18 19:08:35 INFO mapred.JobClient:   Job Counters
08/01/18 19:08:35 INFO mapred.JobClient:     Launched map tasks=3
08/01/18 19:08:35 INFO mapred.JobClient:     Launched reduce tasks=1
08/01/18 19:08:35 INFO mapred.JobClient:     Data-local map tasks=2
08/01/18 19:08:35 INFO mapred.JobClient:   Map-Reduce Framework
08/01/18 19:08:35 INFO mapred.JobClient:     Map input records=1
08/01/18 19:08:35 INFO mapred.JobClient:     Map output records=1
08/01/18 19:08:35 INFO mapred.JobClient:     Map input bytes=2
08/01/18 19:08:35 INFO mapred.JobClient:     Map output bytes=5
08/01/18 19:08:35 INFO mapred.JobClient:     Reduce input groups=1
08/01/18 19:08:35 INFO mapred.JobClient:     Reduce input records=1
08/01/18 19:08:35 INFO mapred.JobClient:     Reduce output records=1
DataLines       Maps    Reduces AvgTime (milliseconds)
1               2       1       926333



On Fri, 18 Jan 2008, Konstantin Shvachko wrote:

Looks like we still have this unsolved mysterious problem:

http://issues.apache.org/jira/browse/HADOOP-1374

Could it be related to HADOOP-1246? Arun?

Thanks,
--Konstantin

Yunhong Gu1 wrote:

Hi,

If someone knows how to fix the problem described below,
please help
me out. Thanks!

I am testing Hadoop on 2-node cluster and the "reduce"
always hangs
at some stage, even if I use different clusters. My OS is Debian
Linux kernel 2.6 (AMD Opteron w/ 4GB Mem). Hadoop verision
is 0.15.2.
Java version is 1.5.0_01-b08.

I simply tried "./bin/hadoop jar hadoop-0.15.2-test.jar
mrbench" and
when the map stage finishes, the reduce stage will hang
somewhere in
the middle, sometimes at 0%. I also tried any other
mapreduce program
I can find in the example jar package but they all hang.

The log file simply print
2008-01-18 15:15:50,831 INFO org.apache.hadoop.mapred.TaskTracker:
task_200801181424_0004_r_000000_0 0.0% reduce > copy >
2008-01-18 15:15:56,841 INFO org.apache.hadoop.mapred.TaskTracker:
task_200801181424_0004_r_000000_0 0.0% reduce > copy >
2008-01-18 15:16:02,850 INFO org.apache.hadoop.mapred.TaskTracker:
task_200801181424_0004_r_000000_0 0.0% reduce > copy >

forever.

The program does work if I start Hadoop only on single node.

Below is my hadoop-site.xml configuration:

<configuration>

<property>
   <name>fs.default.name</name>
   <value>10.0.0.1:60000</value>
</property>

<property>
   <name>mapred.job.tracker</name>
   <value>10.0.0.1:60001</value>
</property>

<property>
   <name>dfs.data.dir</name>
   <value>/raid/hadoop/data</value>
</property>

<property>
   <name>mapred.local.dir</name>
   <value>/raid/hadoop/mapred</value>
</property>

<property>
  <name>hadoop.tmp.dir</name>
  <value>/raid/hadoop/tmp</value>
</property>

<property>
  <name>mapred.child.java.opts</name>
  <value>-Xmx1024m</value>
</property>

<property>
  <name>mapred.tasktracker.tasks.maximum</name>
  <value>4</value>
</property>

<!--
<property>
  <name>mapred.map.tasks</name>
  <value>7</value>
</property>

<property>
  <name>mapred.reduce.tasks</name>
  <value>3</value>
</property>
-->

<property>
  <name>fs.inmemory.size.mb</name>
  <value>200</value>
</property>

<property>
  <name>dfs.block.size</name>
  <value>134217728</value>
</property>

<property>
  <name>io.sort.factor</name>
  <value>100</value>
</property>

<property>
  <name>io.sort.mb</name>
  <value>200</value>
</property>

<property>
  <name>io.file.buffer.size</name>
  <value>131072</value>
</property>

</configuration>






Reply via email to