You can run with a small shell script. You need to override couple of
environment and config variables.
something like :
run_datanode () {
DN=$2
HADOOP_LOG_DIR=logs$DN
HADOOP_PID_DIR=$HADOOP_LOG_DIR
bin/hadoop-daemon.sh $1 datanode \
-Dhadoop.tmp.dir=/some/dir/dfs$DN \
-Ddfs.datanode.address=0.0.0.0:5001$DN \
-Ddfs.datanode.http.address=0.0.0.0:5008$DN \
-Ddfs.datanode.ipc.address=0.0.0.0:5002$DN
}
You can start second datanode like : run_datanode start 2
Pretty useful for testing.
Raghu.
Ajit Ratnaparkhi wrote:
Raghu,
Can you please tell me how to run multiple datanodes on one machine.
thanks,
-Ajit.
On Thu, Feb 26, 2009 at 9:23 AM, Pradeep Fernando <[email protected]>wrote:
Raghu,
I guess you are asking if it would be more convenient if one had access to
a
larger cluster for development.
exactly.....
I have access to many machines and clusters.. but about 99% of my
development happens using single machine for testing. I would guess that
is
true for most of the Hadoop developers.
well this is the answer I was looking for.... :D
seems to be I have enough resources to contribute to this project.
Thanks a lot raghu.
regards,
Pradeep Fernando.