But if we have two master in the master file we have master and secondary node, *both *processes getting started on the two servers listed. Cant we have master and secondary node started seperately on two machines??
On Fri, May 15, 2009 at 9:39 AM, jason hadoop <[email protected]>wrote: > I agree with billy. conf/masters is misleading as the place for secondary > namenodes. > > On Thu, May 14, 2009 at 8:38 PM, Billy Pearson > <[email protected]>wrote: > > > I thank the secondary namenode is set in the masters file in the conf > > folder > > misleading > > > > Billy > > > > > > > > "Rakhi Khatwani" <[email protected]> wrote in message > > news:[email protected]... > > > > Hi, > >> I wanna set up a cluster of 5 nodes in such a way that > >> node1 - master > >> node2 - secondary namenode > >> node3 - slave > >> node4 - slave > >> node5 - slave > >> > >> > >> How do we go about that? > >> there is no property in hadoop-env where i can set the ip-address for > >> secondary name node. > >> > >> if i set node-1 and node-2 in masters, and when we start dfs, in both > the > >> m/cs, the namenode n secondary namenode processes r present. but i think > >> only node1 is active. > >> n my namenode fail over operation fails. > >> > >> ny suggesstions? > >> > >> Regards, > >> Rakhi > >> > >> > > > > > > > -- > Alpha Chapters of my book on Hadoop are available > http://www.apress.com/book/view/9781430219422 > www.prohadoopbook.com a community for Hadoop Professionals >
