Hello David,

thanks a lot. Yet I want java code to submit my application. I do not want to mess with any kind of command line arguments or an executable, neither Java nor Hadoop. I want to write a method that can set up and submit a job to an arbitrary cluster. Something like calling CustomJob.submitJob(ip:port). This would be used by a GUI or another java application to process data. I suspect that the classes Cluster and Job will solve my problem as proposed earlier. The problem is the the missing job.jar, as also described earlier. I will start a new thread describing my problem using a more accurate header.

Thank you,
Martin



On 24.09.2010 20:44, David Rosenstrauch wrote:
On 09/24/2010 01:26 PM, Martin Becker wrote:
Hello David,

This will at best run my MapReduce process on the local Hadoop instance.
What do I do to submit it to a remote Hadoop cluster using Java code?

Martin

$ java -cp <jars> YourApp -libjars <jars> -jt <hostname_of_job_tracker_in_remote_cluster:job_tracker_port_number> -fs <hdfs://hostname_of_name_nod_in_remote_cluster:name_node_port_number> <parms>

DR

Reply via email to