Siddharth, please check 'mapred.local.dir', but I would like advice you check GC logs and OS logs. pay more attention on OS logs. I suspect you start too many threads concurrently, then consumed all OS avaliable resources.
On Thu, Nov 28, 2013 at 9:08 AM, Vinayakumar B <vinayakuma...@huawei.com>wrote: > Hi Siddharth, > > > > Looks like the issue with one of the machine. Or its happening in > different machines also? > > > > I don’t think it’s a problem with JVM heap memory. > > > > Suggest you to check this once, > > http://stackoverflow.com/questions/8384000/java-io-ioexception-error-11 > > > > Thanks and Regards, > > Vinayakumar B > > > > *From:* Siddharth Tiwari [mailto:siddharth.tiw...@live.com] > *Sent:* 28 November 2013 05:50 > *To:* USers Hadoop > *Subject:* RE: Error for larger jobs > > > > Hi Azury > > > > Thanks for response. I have plenty of space on my Disks so that cannot be > the issue. > > > **------------------------** > *Cheers !!!* > *Siddharth Tiwari* > Have a refreshing day !!! > *"Every duty is holy, and devotion to duty is the highest form of worship > of God.” * > *"Maybe other people will try to limit me but I don't limit myself"* > > ------------------------------ > > Date: Thu, 28 Nov 2013 08:10:06 +0800 > Subject: Re: Error for larger jobs > From: azury...@gmail.com > To: user@hadoop.apache.org > > Your disk is full from the log. > > On 2013-11-28 5:27 AM, "Siddharth Tiwari" <siddharth.tiw...@live.com> > wrote: > > Hi Team > > > > I am getting following strange error, can you point me to the possible > reason. > > I have set heap size to 4GB but still getting it. please help > > > > *syslog logs* > > 2013-11-27 19:01:50,678 WARN org.apache.hadoop.util.NativeCodeLoader: > Unable to load native-hadoop library for your platform... using > builtin-java classes where applicable > > 2013-11-27 19:01:51,051 WARN mapreduce.Counters: Group > org.apache.hadoop.mapred.Task$Counter is deprecated. Use > org.apache.hadoop.mapreduce.TaskCounter instead > > 2013-11-27 19:01:51,539 WARN org.apache.hadoop.conf.Configuration: > session.id is deprecated. Instead, use dfs.metrics.session-id > > 2013-11-27 19:01:51,540 INFO org.apache.hadoop.metrics.jvm.JvmMetrics: > Initializing JVM Metrics with processName=MAP, sessionId= > > 2013-11-27 19:01:51,867 INFO org.apache.hadoop.util.ProcessTree: setsid > exited with exit code 0 > > 2013-11-27 19:01:51,870 INFO org.apache.hadoop.mapred.Task: Using > ResourceCalculatorPlugin : > org.apache.hadoop.util.LinuxResourceCalculatorPlugin@4a0bd13d > > 2013-11-27 19:01:52,217 INFO org.apache.hadoop.mapred.MapTask: Processing > split: > org.apache.hadoop.examples.terasort.TeraGen$RangeInputFormat$RangeInputSplit@6c30aec7 > > 2013-11-27 19:01:52,222 WARN mapreduce.Counters: Counter name > MAP_INPUT_BYTES is deprecated. Use FileInputFormatCounters as group name > and BYTES_READ as counter name instead > > 2013-11-27 19:01:52,226 INFO org.apache.hadoop.mapred.MapTask: > numReduceTasks: 0 > > 2013-11-27 19:01:52,250 ERROR > org.apache.hadoop.security.UserGroupInformation: PriviledgedActionException > as:hadoop (auth:SIMPLE) cause:java.io.IOException: Cannot run program > "chmod": error=11, Resource temporarily unavailable > > 2013-11-27 19:01:52,250 WARN org.apache.hadoop.mapred.Child: Error running > child > > java.io.IOException: Cannot run program "chmod": error=11, Resource > temporarily unavailable > > at java.lang.ProcessBuilder.start(ProcessBuilder.java:1041) > > at org.apache.hadoop.util.Shell.runCommand(Shell.java:206) > > at org.apache.hadoop.util.Shell.run(Shell.java:188) > > at > org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:381) > > at org.apache.hadoop.util.Shell.execCommand(Shell.java:467) > > at org.apache.hadoop.util.Shell.execCommand(Shell.java:450) > > at > org.apache.hadoop.fs.RawLocalFileSystem.execCommand(RawLocalFileSystem.java:593) > > at > org.apache.hadoop.fs.RawLocalFileSystem.setPermission(RawLocalFileSystem.java:584) > > at > org.apache.hadoop.io.SecureIOUtils.insecureCreateForWrite(SecureIOUtils.java:146) > > at > org.apache.hadoop.io.SecureIOUtils.createForWrite(SecureIOUtils.java:168) > > at > org.apache.hadoop.mapred.TaskLog.writeToIndexFile(TaskLog.java:310) > > at org.apache.hadoop.mapred.TaskLog.syncLogs(TaskLog.java:383) > > at org.apache.hadoop.mapred.Child$4.run(Child.java:270) > > at java.security.AccessController.doPrivileged(Native Method) > > at javax.security.auth.Subject.doAs(Subject.java:415) > > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1408) > > at org.apache.hadoop.mapred.Child.main(Child.java:262) > > Caused by: java.io.IOException: error=11, Resource temporarily unavailable > > at java.lang.UNIXProcess.forkAndExec(Native Method) > > at java.lang.UNIXProcess.<init>(UNIXProcess.java:135) > > at java.lang.ProcessImpl.start(ProcessImpl.java:130) > > at java.lang.ProcessBuilder.start(ProcessBuilder.java:1022) > > ... 16 more > > 2013-11-27 19:01:52,256 INFO org.apache.hadoop.mapred.Task: Runnning > cleanup for the task > > > > **------------------------** > *Cheers !!!* > *Siddharth Tiwari* > Have a refreshing day !!! > *"Every duty is holy, and devotion to duty is the highest form of worship > of God.” * > *"Maybe other people will try to limit me but I don't limit myself"* >