[
https://issues.apache.org/jira/browse/MAPREDUCE-5901?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Vinod Kumar Vavilapalli resolved MAPREDUCE-5901.
------------------------------------------------
Resolution: Won't Fix
Release Note: (was: 2.4)
Folks have reported working of multiple node clusters.
In any case, please use the user mailing lists for debugging such issues. JIRA
is for tracking bugs.
> Hadoop 2.4 Java execution issue: remotely submission jobs fail
> ---------------------------------------------------------------
>
> Key: MAPREDUCE-5901
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-5901
> Project: Hadoop Map/Reduce
> Issue Type: New Feature
> Environment: java, hadoop v. 2.4
> Reporter: michele
>
> I have installed Hadoop 2.4 on remote machine in Single-Mode setting. From
> another machine (client) I run a Java application that submit a job to a
> remote Hadoop machine (cluster), I have used the attached code. The problem
> is that the real execution of the map process is run on my local machine
> (client) not on the cluster machine.
> JobConf job = new JobConf(SOF.class);
> job.setJobName("SIM-"+sim_id);
> System.setProperty("HADOOP_USER_NAME", "hadoop");
> FileInputFormat.addInputPath(job,new
> Path("hdfs://cluster_ip:port"+USERS_HOME+user+"/SIM-"+sim_id+"/"+INPUT_FOLDER_HOME+"/input.tmp")/*new_inputs_path*/);
> FileOutputFormat.setOutputPath(job, new
> Path("hdfs://cluster_ip:port"+USERS_HOME+user+"/SIM-"+sim_id+"/"+OUTPUT_FOLDER_HOME));
> job.set("jar.work.directory",
> "hdfs://cluster_ip:port"+SOF.USERS_HOME+user+"/SIM-"+sim_id+"/flockers.jar");
> job.setMapperClass(Mapper.class);
> job.setReducerClass(Reducer.class);
> job.setOutputKeyClass(org.apache.hadoop.io.Text.class);
> job.setOutputValueClass(org.apache.hadoop.io.Text.class);
> job.set("mapred.job.tracker", "cluster_ip:port");
> job.set("fs.default.name", "hdfs://cluster_ip:port");
> job.set("hadoop.job.ugi", "hadoop,hadoop");
> job.set("user", "hadoop");
> try {
> JobClient jobc=new JobClient(job);
> System.out.println(jobc+" "+job);
> RunningJob runjob;
> runjob = jobc.submitJob(job);
> System.out.println(runjob);
> System.out.println("VM "+Inet4Address.getLocalHost());
> while(runjob.getJobStatus().equals(JobStatus.SUCCEEDED)){}
> } catch (Exception e) {
> // TODO Auto-generated catch block
> e.printStackTrace();
> }
> }
> I have tried to set up correctly hadoop using the following mapred-site.xml:
> <configuration>
> <property>
> <name>mapred.job.tracker</name>
> <value>cluster_ip:port</value>
> </property>
> <property>
> <name>mapreduce.framework.name</name>
> <value>yarn</value>
> </property>
> </configuration>
--
This message was sent by Atlassian JIRA
(v6.2#6252)