Hi,

I saw that pretty much after sending the email. I verified the properties
file and it has all the correct properties even the
mapred.framework.nameis set to yarn. I am unable to figure out what is
the cause and why it is
connecting to local FS.

Using the same configuration file I am able to run my WordCount MRV1
example but not the code that I have written for a usecase.

Thanks
Divye Sheth


On Tue, Apr 8, 2014 at 6:12 PM, Kavali, Devaraj <devaraj.kav...@intel.com>wrote:

>  As per the given exception stack trace, it is trying to use local file
> system. Can you check whether you have configured the file system
> configurations with HDFS?
>
>
>
> Thanks
>
> Devaraj K
>
>
>
> *From:* divye sheth [mailto:divs.sh...@gmail.com]
> *Sent:* Tuesday, April 08, 2014 5:37 PM
> *To:* user@hadoop.apache.org
> *Subject:* Running MRV1 code on YARN
>
>
>
> Hi,
>
>
>
> I have installed Hadoop 2.2.0 along with YARN and am trying to submit a
> MRV1 job already written to YARN.
>
>
>
> The job does not even submit and it prints the following stack trace on
> console:
>
>
>
> 2014-04-08 16:56:11 UserGroupInformation [ERROR]
> PriviledgedActionException as:eureka (auth:SIMPLE)
> cause:org.apache.hadoop.util.Shell$ExitCodeException: chmod: cannot access
> `/user/eureka54695942/.staging/job_local54695942_0001': No such file or
> directory
>
>
>
> Exception in thread "main" org.apache.hadoop.util.Shell$ExitCodeException:
> chmod: cannot access
> `/user/eureka54695942/.staging/job_local54695942_0001': No such file or
> directory
>
>
>
>         at org.apache.hadoop.util.Shell.runCommand(Shell.java:261)
>
>         at org.apache.hadoop.util.Shell.run(Shell.java:188)
>
>         at
> org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:381)
>
>         at org.apache.hadoop.util.Shell.execCommand(Shell.java:467)
>
>         at org.apache.hadoop.util.Shell.execCommand(Shell.java:450)
>
>         at
> org.apache.hadoop.fs.RawLocalFileSystem.execCommand(RawLocalFileSystem.java:593)
>
>         at
> org.apache.hadoop.fs.RawLocalFileSystem.setPermission(RawLocalFileSystem.java:584)
>
>         at
> org.apache.hadoop.fs.FilterFileSystem.setPermission(FilterFileSystem.java:427)
>
>         at org.apache.hadoop.fs.FileSystem.mkdirs(FileSystem.java:579)
>
>         at
> org.apache.hadoop.mapred.JobClient.copyAndConfigureFiles(JobClient.java:786)
>
>         at
> org.apache.hadoop.mapred.JobClient.copyAndConfigureFiles(JobClient.java:746)
>
>         at
> org.apache.hadoop.mapred.JobClient.access$400(JobClient.java:177)
>
>         at org.apache.hadoop.mapred.JobClient$2.run(JobClient.java:963)
>
>         at org.apache.hadoop.mapred.JobClient$2.run(JobClient.java:948)
>
>         at java.security.AccessController.doPrivileged(Native Method)
>
>         at javax.security.auth.Subject.doAs(Subject.java:396)
>
>         at
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1408)
>
>         at
> org.apache.hadoop.mapred.JobClient.submitJobInternal(JobClient.java:948)
>
>         at org.apache.hadoop.mapreduce.Job.submit(Job.java:566)
>
>
>
> My question here is if you notice the staging location which it is trying
> to clean I do not have any such user in the /user directory in hdfs. It
> somehow appends the jobId to the username and creates staging area there.
> Any reason for this? Please let me know what am I doing wrong. How can I
> make sure it goes to the user that I have created i.e. eureka and not
> eureka$JOBID.
>
>
>
> I am using CDH4.
>
>
>
> Thanks
>
> Divye Sheth
>

Reply via email to