Actually I'm trying to do something like combining multiple namenodes so that they present themselves to clients as a single namespace, implementing basic namenode functionalities.
在 2013年4月18日星期四,Chris Embree 写道: > Glad you got this working... can you explain your use case a little? I'm > trying to understand why you might want to do that. > > > On Thu, Apr 18, 2013 at 11:29 AM, Lixiang Ao > <aolixi...@gmail.com<javascript:_e({}, 'cvml', 'aolixi...@gmail.com');> > > wrote: > >> I modified sbin/hadoop-daemon.sh, where HADOOP_PID_DIR is set. It works! >> Everything looks fine now. >> >> Seems direct command "hdfs namenode" gives a better sense of control :) >> >> Thanks a lot. >> >> 在 2013年4月18日星期四,Harsh J 写道: >> >> Yes you can but if you want the scripts to work, you should have them >>> use a different PID directory (I think its called HADOOP_PID_DIR) >>> every time you invoke them. >>> >>> I instead prefer to start the daemons up via their direct command such >>> as "hdfs namenode" and so and move them to the background, with a >>> redirect for logging. >>> >>> On Thu, Apr 18, 2013 at 2:34 PM, Lixiang Ao <aolixi...@gmail.com> wrote: >>> > Hi all, >>> > >>> > Can I run mutiple HDFS instances, that is, n seperate namenodes and n >>> > datanodes, on a single machine? >>> > >>> > I've modified core-site.xml and hdfs-site.xml to avoid port and file >>> > conflicting between HDFSes, but when I started the second HDFS, I got >>> the >>> > errors: >>> > >>> > Starting namenodes on [localhost] >>> > localhost: namenode running as process 20544. Stop it first. >>> > localhost: datanode running as process 20786. Stop it first. >>> > Starting secondary namenodes [0.0.0.0] >>> > 0.0.0.0: secondarynamenode running as process 21074. Stop it first. >>> > >>> > Is there a way to solve this? >>> > Thank you in advance, >>> > >>> > Lixiang Ao >>> >>> >>> >>> -- >>> Harsh J >>> >> >