Siddharth,
please check 'mapred.local.dir', but I would like advice you check GC logs
and OS logs. pay more attention on OS logs. I suspect you start too many
threads concurrently, then consumed all OS avaliable resources.



On Thu, Nov 28, 2013 at 9:08 AM, Vinayakumar B <vinayakuma...@huawei.com>wrote:

>  Hi Siddharth,
>
>
>
> Looks like the issue with one of the machine.  Or its happening in
> different machines also?
>
>
>
> I don’t think it’s a problem with JVM heap memory.
>
>
>
> Suggest you to check this once,
>
> http://stackoverflow.com/questions/8384000/java-io-ioexception-error-11
>
>
>
> Thanks and Regards,
>
> Vinayakumar B
>
>
>
> *From:* Siddharth Tiwari [mailto:siddharth.tiw...@live.com]
> *Sent:* 28 November 2013 05:50
> *To:* USers Hadoop
> *Subject:* RE: Error for larger jobs
>
>
>
> Hi Azury
>
>
>
> Thanks for response. I have plenty of space on my Disks so that cannot be
> the issue.
>
>
> **------------------------**
> *Cheers !!!*
> *Siddharth Tiwari*
> Have a refreshing day !!!
> *"Every duty is holy, and devotion to duty is the highest form of worship
> of God.” *
> *"Maybe other people will try to limit me but I don't limit myself"*
>
>   ------------------------------
>
> Date: Thu, 28 Nov 2013 08:10:06 +0800
> Subject: Re: Error for larger jobs
> From: azury...@gmail.com
> To: user@hadoop.apache.org
>
> Your disk is full from the log.
>
> On 2013-11-28 5:27 AM, "Siddharth Tiwari" <siddharth.tiw...@live.com>
> wrote:
>
> Hi Team
>
>
>
> I am getting following strange error, can you point me to the possible
> reason.
>
> I have set heap size to 4GB but still getting it. please help
>
>
>
> *syslog logs*
>
> 2013-11-27 19:01:50,678 WARN org.apache.hadoop.util.NativeCodeLoader:
> Unable to load native-hadoop library for your platform... using
> builtin-java classes where applicable
>
> 2013-11-27 19:01:51,051 WARN mapreduce.Counters: Group
> org.apache.hadoop.mapred.Task$Counter is deprecated. Use
> org.apache.hadoop.mapreduce.TaskCounter instead
>
> 2013-11-27 19:01:51,539 WARN org.apache.hadoop.conf.Configuration:
> session.id is deprecated. Instead, use dfs.metrics.session-id
>
> 2013-11-27 19:01:51,540 INFO org.apache.hadoop.metrics.jvm.JvmMetrics:
> Initializing JVM Metrics with processName=MAP, sessionId=
>
> 2013-11-27 19:01:51,867 INFO org.apache.hadoop.util.ProcessTree: setsid
> exited with exit code 0
>
> 2013-11-27 19:01:51,870 INFO org.apache.hadoop.mapred.Task:  Using
> ResourceCalculatorPlugin :
> org.apache.hadoop.util.LinuxResourceCalculatorPlugin@4a0bd13d
>
> 2013-11-27 19:01:52,217 INFO org.apache.hadoop.mapred.MapTask: Processing
> split:
> org.apache.hadoop.examples.terasort.TeraGen$RangeInputFormat$RangeInputSplit@6c30aec7
>
> 2013-11-27 19:01:52,222 WARN mapreduce.Counters: Counter name
> MAP_INPUT_BYTES is deprecated. Use FileInputFormatCounters as group name
> and  BYTES_READ as counter name instead
>
> 2013-11-27 19:01:52,226 INFO org.apache.hadoop.mapred.MapTask:
> numReduceTasks: 0
>
> 2013-11-27 19:01:52,250 ERROR
> org.apache.hadoop.security.UserGroupInformation: PriviledgedActionException
> as:hadoop (auth:SIMPLE) cause:java.io.IOException: Cannot run program
> "chmod": error=11, Resource temporarily unavailable
>
> 2013-11-27 19:01:52,250 WARN org.apache.hadoop.mapred.Child: Error running
> child
>
> java.io.IOException: Cannot run program "chmod": error=11, Resource
> temporarily unavailable
>
>         at java.lang.ProcessBuilder.start(ProcessBuilder.java:1041)
>
>         at org.apache.hadoop.util.Shell.runCommand(Shell.java:206)
>
>         at org.apache.hadoop.util.Shell.run(Shell.java:188)
>
>         at
> org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:381)
>
>         at org.apache.hadoop.util.Shell.execCommand(Shell.java:467)
>
>         at org.apache.hadoop.util.Shell.execCommand(Shell.java:450)
>
>         at
> org.apache.hadoop.fs.RawLocalFileSystem.execCommand(RawLocalFileSystem.java:593)
>
>         at
> org.apache.hadoop.fs.RawLocalFileSystem.setPermission(RawLocalFileSystem.java:584)
>
>         at
> org.apache.hadoop.io.SecureIOUtils.insecureCreateForWrite(SecureIOUtils.java:146)
>
>         at
> org.apache.hadoop.io.SecureIOUtils.createForWrite(SecureIOUtils.java:168)
>
>         at
> org.apache.hadoop.mapred.TaskLog.writeToIndexFile(TaskLog.java:310)
>
>         at org.apache.hadoop.mapred.TaskLog.syncLogs(TaskLog.java:383)
>
>         at org.apache.hadoop.mapred.Child$4.run(Child.java:270)
>
>         at java.security.AccessController.doPrivileged(Native Method)
>
>         at javax.security.auth.Subject.doAs(Subject.java:415)
>
>         at
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1408)
>
>         at org.apache.hadoop.mapred.Child.main(Child.java:262)
>
> Caused by: java.io.IOException: error=11, Resource temporarily unavailable
>
>         at java.lang.UNIXProcess.forkAndExec(Native Method)
>
>         at java.lang.UNIXProcess.<init>(UNIXProcess.java:135)
>
>         at java.lang.ProcessImpl.start(ProcessImpl.java:130)
>
>         at java.lang.ProcessBuilder.start(ProcessBuilder.java:1022)
>
>         ... 16 more
>
> 2013-11-27 19:01:52,256 INFO org.apache.hadoop.mapred.Task: Runnning
> cleanup for the task
>
>
>
> **------------------------**
> *Cheers !!!*
> *Siddharth Tiwari*
> Have a refreshing day !!!
> *"Every duty is holy, and devotion to duty is the highest form of worship
> of God.” *
> *"Maybe other people will try to limit me but I don't limit myself"*
>

Reply via email to