[ 
https://issues.apache.org/jira/browse/MAPREDUCE-6438?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Raghvendra Singh updated MAPREDUCE-6438:
----------------------------------------
    Attachment: yarn-site.xml
                mapred-site.xml
                hdfs-site.xml
                core-site.xml

> mapreduce fails with job.jar does not exist
> -------------------------------------------
>
>                 Key: MAPREDUCE-6438
>                 URL: https://issues.apache.org/jira/browse/MAPREDUCE-6438
>             Project: Hadoop Map/Reduce
>          Issue Type: Bug
>          Components: mrv2
>            Reporter: Raghvendra Singh
>         Attachments: core-site.xml, hdfs-site.xml, mapred-site.xml, 
> yarn-site.xml
>
>
> I have a hortonworks distribution(2.2.6.0-2800) of hadoop which runs 
> mapreduce job based on yarn, and i have a simple map reduce job which reads 
> compressed data files from hdfs, does some processing over it and then this 
> data is saved in hbase with bulk load
> Here is my program that does it
> {code}
>         final Configuration hadoopConfiguration = new Configuration();
>         configuration.set(“yarn.resourcemanager.address”, “XXXXXX”);
>         configuration.set(“yarn.resourcemanager.scheduler.address”, “XXXXXX”);
>         configuration.set("mapreduce.framework.name", "yarn”);
>         configuration.set("mapreduce.jobtracker.staging.root.dir", 
> “XXXXXXXX”);
>         final Job job = Job.getInstance(hadoopConfiguration, "migration");
>         job.setJarByClass(BlitzService.class);
>         job.setMapperClass(DataMigrationMapper.class);
>         job.setMapOutputKeyClass(ImmutableBytesWritable.class);
>         job.setMapOutputValueClass(KeyValue.class);
>         job.setReducerClass(DataMigrationReducer.class);
>         job.setCombinerClass(DataMigrationReducer.class);
>         HFileOutputFormat2.configureIncrementalLoad(job, hTable);
>         FileInputFormat.setInputPaths(job, filesToProcess.toArray(new 
> Path[filesToProcess.size()]));
>         HFileOutputFormat2.setOutputPath(job, new Path(SOME PATH));
>         job.waitForCompletion(true);
> {code}
> This should be a very simple thing to run but i am facing this exception 
> while running the job
> {code}
>     INFO  [2015-07-23 23:53:20,222] org.apache.hadoop.yarn.client.RMProxy: 
> Connecting to ResourceManager at /172.30.0.147:8032
>     WARN  [2015-07-23 23:53:20,383] org.apache.hadoop.mapreduce.JobSubmitter: 
> Hadoop command-line option parsing not performed. Implement the Tool 
> interface and execute your application with ToolRunner to remedy this.
>     INFO  [2015-07-23 23:53:20,492] 
> org.apache.hadoop.mapreduce.lib.input.FileInputFormat: Total input paths to 
> process : 16
>     INFO  [2015-07-23 23:53:20,561] org.apache.hadoop.mapreduce.JobSubmitter: 
> number of splits:16
>     INFO  [2015-07-23 23:53:20,719] org.apache.hadoop.mapreduce.JobSubmitter: 
> Submitting tokens for job: job_1437695344326_0002
>     INFO  [2015-07-23 23:53:20,842] 
> org.apache.hadoop.yarn.client.api.impl.YarnClientImpl: Submitted application 
> application_1437695344326_0002
>     INFO  [2015-07-23 23:53:20,867] org.apache.hadoop.mapreduce.Job: The url 
> to track the job: 
> http://ip-172-30-0-147.us-west-2.compute.internal:8088/proxy/application_1437695344326_0002/
>     INFO  [2015-07-23 23:53:20,868] org.apache.hadoop.mapreduce.Job: Running 
> job: job_1437695344326_0002
>     INFO  [2015-07-23 23:53:35,994] org.apache.hadoop.mapreduce.Job: Job 
> job_1437695344326_0002 running in uber mode : false
>     INFO  [2015-07-23 23:53:35,995] org.apache.hadoop.mapreduce.Job:  map 0% 
> reduce 0%
>     INFO  [2015-07-23 23:53:43,053] org.apache.hadoop.mapreduce.Job: Task Id 
> : attempt_1437695344326_0002_m_000001_1000, Status : FAILED
>     File 
> file:/tmp/hadoop-yarn/staging/root/.staging/job_1437695344326_0002/job.jar 
> does not exist
>     java.io.FileNotFoundException: File 
> file:/tmp/hadoop-yarn/staging/root/.staging/job_1437695344326_0002/job.jar 
> does not exist
>       at 
> org.apache.hadoop.fs.RawLocalFileSystem.deprecatedGetFileStatus(RawLocalFileSystem.java:608)
>       at 
> org.apache.hadoop.fs.RawLocalFileSystem.getFileLinkStatusInternal(RawLocalFileSystem.java:821)
>       at 
> org.apache.hadoop.fs.RawLocalFileSystem.getFileStatus(RawLocalFileSystem.java:598)
>       at 
> org.apache.hadoop.fs.FilterFileSystem.getFileStatus(FilterFileSystem.java:414)
>       at org.apache.hadoop.yarn.util.FSDownload.copy(FSDownload.java:251)
>       at org.apache.hadoop.yarn.util.FSDownload.access$000(FSDownload.java:61)
>       at org.apache.hadoop.yarn.util.FSDownload$2.run(FSDownload.java:359)
>       at org.apache.hadoop.yarn.util.FSDownload$2.run(FSDownload.java:357)
>       at java.security.AccessController.doPrivileged(Native Method)
>       at javax.security.auth.Subject.doAs(Subject.java:415)
>       at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628)
>       at org.apache.hadoop.yarn.util.FSDownload.call(FSDownload.java:356)
>       at org.apache.hadoop.yarn.util.FSDownload.call(FSDownload.java:60)
>       at java.util.concurrent.FutureTask.run(FutureTask.java:262)
>       at 
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
>       at java.util.concurrent.FutureTask.run(FutureTask.java:262)
>       at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>       at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>       at java.lang.Thread.run(Thread.java:745)
>     
>     
>     INFO  [2015-07-23 23:53:44,075] org.apache.hadoop.mapreduce.Job: Task Id 
> : attempt_1437695344326_0002_m_000002_1000, Status : FAILED
>     File 
> file:/tmp/hadoop-yarn/staging/root/.staging/job_1437695344326_0002/job.jar 
> does not exist
>     java.io.FileNotFoundException: File 
> file:/tmp/hadoop-yarn/staging/root/.staging/job_1437695344326_0002/job.jar 
> does not exist
>       at 
> org.apache.hadoop.fs.RawLocalFileSystem.deprecatedGetFileStatus(RawLocalFileSystem.java:608)
>       at 
> org.apache.hadoop.fs.RawLocalFileSystem.getFileLinkStatusInternal(RawLocalFileSystem.java:821)
>       at 
> org.apache.hadoop.fs.RawLocalFileSystem.getFileStatus(RawLocalFileSystem.java:598)
>       at 
> org.apache.hadoop.fs.FilterFileSystem.getFileStatus(FilterFileSystem.java:414)
>       at org.apache.hadoop.yarn.util.FSDownload.copy(FSDownload.java:251)
>       at org.apache.hadoop.yarn.util.FSDownload.access$000(FSDownload.java:61)
>       at org.apache.hadoop.yarn.util.FSDownload$2.run(FSDownload.java:359)
>       at org.apache.hadoop.yarn.util.FSDownload$2.run(FSDownload.java:357)
>       at java.security.AccessController.doPrivileged(Native Method)
>       at javax.security.auth.Subject.doAs(Subject.java:415)
>       at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628)
>       at org.apache.hadoop.yarn.util.FSDownload.call(FSDownload.java:356)
>       at org.apache.hadoop.yarn.util.FSDownload.call(FSDownload.java:60)
>       at java.util.concurrent.FutureTask.run(FutureTask.java:262)
>       at 
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
>       at java.util.concurrent.FutureTask.run(FutureTask.java:262)
>       at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>       at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>       at java.lang.Thread.run(Thread.java:745)
> {code}
> Also attached are the core-site.xml, mapred-site.xml, hdfs-site.xml and 
> yarn-site.xml



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to