You can run with a small shell script. You need to override couple of environment and config variables.

something like :

run_datanode () {
        DN=$2
        HADOOP_LOG_DIR=logs$DN
        HADOOP_PID_DIR=$HADOOP_LOG_DIR
        bin/hadoop-daemon.sh $1 datanode \
          -Dhadoop.tmp.dir=/some/dir/dfs$DN \
          -Ddfs.datanode.address=0.0.0.0:5001$DN \
          -Ddfs.datanode.http.address=0.0.0.0:5008$DN \
          -Ddfs.datanode.ipc.address=0.0.0.0:5002$DN
}

You can start second datanode like : run_datanode start 2

Pretty useful for testing.

Raghu.

Ajit Ratnaparkhi wrote:
Raghu,

Can you please tell me how to run multiple datanodes on one machine.

thanks,
-Ajit.

On Thu, Feb 26, 2009 at 9:23 AM, Pradeep Fernando <[email protected]>wrote:

Raghu,

I guess you are asking if it would be more convenient if one had access to
a
larger cluster for development.

exactly.....

 I have access to many machines and clusters.. but about 99% of my
development happens using single machine for testing. I would guess that
is
true for most of the Hadoop developers.

well this is the answer I was looking for....  :D
seems to be I have enough resources to contribute to this project.
Thanks a lot raghu.

regards,
Pradeep Fernando.



Reply via email to