When running a job standalone (typically for debugging), it suffices for me
to simply add -Dmapred.job.tracker=local to the program command line.

If you simply want to run the program on a single node, then you can just
stop the other node.  Using local storage with distributed programs is not
recommended.


On 11/1/07 12:35 PM, "Jim the Standing Bear" <[EMAIL PROTECTED]> wrote:

> Hello,
> 
> I am in need of some clarifications on how to run a hadoop job locally.
> 
> The cluster was originally set up to have two nodes, where one of them
> also acts as the master node and job tracker.
> 
> According to the wiki, I can run a job locally by altering
> "mapred.job.tracker" and "fs.default.name" properties to "local" in
> hadoop-site.xml.  But when I start the server, it stack dumped:
> 
> localhost: starting secondarynamenode, logging to /home/blahblahblah
> localhost: Exception in thread "main" java.lang.RuntimeException: Not
> a host:port pair: local
> 
> Apparently it didn't like the value "local"?
> 
> Also, the wiki noted that all these XML configuration files should be
> included somewhere in the class path to the job, does it mean I need
> to include the XMLs as I do jars?
> 
> Thank
> 
> -- Jim

Reply via email to