Rather than memory problem, it was a disk problem. I made more free spaces and 
it fixed


 
Regards,
Mahmood



On Saturday, March 22, 2014 8:58 PM, Mahmood Naderan <nt_mahm...@yahoo.com> 
wrote:
 
Really stuck at this step. I have test with smaller data set and it works. Now 
I am using wikipedia articles (46GB) with 600 chunks (each 64MB)

I have set number of mappers and reducers to 1 to ensure consistency and I am 
running on a local node. Why reducer doesn't report anything within 600 
seconds??


14/03/22 15:00:51 INFO mapred.JobClient:  map 15% reduce 5%
14/03/22 15:18:43 INFO mapred.JobClient:  map 16% reduce 5%
14/03/22 15:46:38 INFO mapred.JobClient: Task Id : 
attempt_201403212248_0002_m_000118_0, Status : FAILED
Task attempt_201403212248_0002_m_000118_0 failed to report status for 600 
seconds. Killing!
14/03/22 15:48:54 INFO mapred.JobClient:  map 17% reduce 5%
14/03/22 16:06:32 INFO mapred.JobClient:  map 18% reduce 5%
14/03/22
 16:07:08 INFO mapred.JobClient:  map 18% reduce 6%
14/03/22 16:24:09 INFO mapred.JobClient:  map 19% reduce 6%
14/03/22 16:41:58 INFO mapred.JobClient:  map 20% reduce 6%
14/03/22 16:55:13 INFO mapred.JobClient: Task Id : 
attempt_201403212248_0002_r_000000_0, Status : FAILED
java.io.IOException: Task: attempt_201403212248_0002_r_000000_0 - The reduce 
copier failed
    at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:390)
    at org.apache.hadoop.mapred.Child$4.run(Child.java:255)
    at java.security.AccessController.doPrivileged(Native Method)
    at javax.security.auth.Subject.doAs(Subject.java:415)
    at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1190)
    at org.apache.hadoop.mapred.Child.main(Child.java:249)
Caused by:
 org.apache.hadoop.util.DiskChecker$DiskErrorException: Could not find any 
valid local directory for 
file:/tmp/hadoop-hadoop/mapred/local/taskTracker/hadoop/jobcache/job_201403212248_0002/attempt_201403212248_0002_r_000000_0/output/map_107.out
    at 
org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.getLocalPathForWrite(LocalDirAllocator.java:381)
    at 
org.apache.hadoop.fs.LocalDirAllocator.getLocalPathForWrite(LocalDirAllocator.java:146)
    at 
org.apache.hadoop.fs.LocalDirAllocator.getLocalPathForWrite(LocalDirAllocator.java:127)
    at 
org.apache.hadoop.mapred.ReduceTask$ReduceCopier$LocalFSMerger.run(ReduceTask.java:2690)

attempt_201403212248_0002_r_000000_0: log4j:WARN No appenders could be found 
for logger (org.apache.hadoop.mapred.Task).
attempt_201403212248_0002_r_000000_0: log4j:WARN Please initialize the log4j 
system properly.
14/03/22 16:55:15 INFO
 mapred.JobClient:  map 20% reduce 0%
14/03/22 16:55:34 INFO mapred.JobClient:  map 20% reduce 1%





 
Regards,
Mahmood



On Saturday, March 22, 2014 10:27 AM, Mahmood Naderan <nt_mahm...@yahoo.com> 
wrote:
 
Again I got the same error and it says

The reducer copier failed
...
could not find any valid local directory for file 
/tmp/hadoop-hadoop/....map_150.out

Searching the web shows that I have to clean up the /tmp/hadoop-hadoop folder 
but the total size of this folder is 800KB with 1100 files. Does that really 
matter?


 
Regards,
Mahmood



On Friday, March 21, 2014 3:52 PM, Mahmood Naderan <nt_mahm...@yahoo.com> wrote:
 
OK it seems that there was a "free disk space" issue.
I made more spaces and running again.


 
Regards,
Mahmood



On Friday, March 21, 2014 11:43 AM, shashwat shriparv 
<dwivedishash...@gmail.com> wrote:
 
​Check if the tmp dir, hdfs remaining or log directory are getting filled up 
while this job runs..​

On Fri, Mar 21, 2014 at 12:11 PM, Mahmood Naderan <nt_mahm...@yahoo.com> wrote:

that imply a *retry* process? Or I have to be wo

​



Warm Regards_∞_
Shashwat Shriparv

Reply via email to