You have to examine the datanode log files the namenode does not start the datanodes, the start script does. The name node passively waits for the datanodes to connect to it.
On Thu, May 14, 2009 at 6:43 PM, Pankil Doshi <forpan...@gmail.com> wrote: > Hello Everyone, > > Actually I had a cluster which was up. > > But i stopped the cluster as i wanted to format it.But cant start it back. > > 1)when i give "start-dfs.sh" I get following on screen > > starting namenode, logging to > /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-namenode-hadoopmaster.out > slave1.local: starting datanode, logging to > /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave1.out > slave3.local: starting datanode, logging to > /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave3.out > slave4.local: starting datanode, logging to > /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave4.out > slave2.local: starting datanode, logging to > /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave2.out > slave5.local: starting datanode, logging to > /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave5.out > slave6.local: starting datanode, logging to > /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave6.out > slave9.local: starting datanode, logging to > /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave9.out > slave8.local: starting datanode, logging to > /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave8.out > slave7.local: starting datanode, logging to > /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave7.out > slave10.local: starting datanode, logging to > /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave10.out > hadoopmaster.local: starting secondarynamenode, logging to > > /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-secondarynamenode-hadoopmaster.out > > > 2) from log file named "hadoop-hadoop-namenode-hadoopmaster.log" I get > following > > > > 2009-05-14 20:28:23,515 INFO org.apache.hadoop.dfs.NameNode: STARTUP_MSG: > /************************************************************ > STARTUP_MSG: Starting NameNode > STARTUP_MSG: host = hadoopmaster/127.0.0.1 > STARTUP_MSG: args = [] > STARTUP_MSG: version = 0.18.3 > STARTUP_MSG: build = > https://svn.apache.org/repos/asf/hadoop/core/branches/branch-0.18 -r > 736250; > compiled by 'ndaley' on Thu Jan 22 23:12:08 UTC 2009 > ************************************************************/ > 2009-05-14 20:28:23,717 INFO org.apache.hadoop.ipc.metrics.RpcMetrics: > Initializing RPC Metrics with hostName=NameNode, port=9000 > 2009-05-14 20:28:23,728 INFO org.apache.hadoop.dfs.NameNode: Namenode up > at: > hadoopmaster.local/192.168.0.1:9000 > 2009-05-14 20:28:23,733 INFO org.apache.hadoop.metrics.jvm.JvmMetrics: > Initializing JVM Metrics with processName=NameNode, sessionId=null > 2009-05-14 20:28:23,743 INFO org.apache.hadoop.dfs.NameNodeMetrics: > Initializing NameNodeMeterics using context > object:org.apache.hadoop.metrics.spi.NullContext > 2009-05-14 20:28:23,856 INFO org.apache.hadoop.fs.FSNamesystem: > > fsOwner=hadoop,hadoop,adm,dialout,fax,cdrom,floppy,tape,audio,dip,video,plugdev,fuse,lpadmin,admin,sambashare > 2009-05-14 20:28:23,856 INFO org.apache.hadoop.fs.FSNamesystem: > supergroup=supergroup > 2009-05-14 20:28:23,856 INFO org.apache.hadoop.fs.FSNamesystem: > isPermissionEnabled=true > 2009-05-14 20:28:23,883 INFO org.apache.hadoop.dfs.FSNamesystemMetrics: > Initializing FSNamesystemMeterics using context > object:org.apache.hadoop.metrics.spi.NullContext > 2009-05-14 20:28:23,885 INFO org.apache.hadoop.fs.FSNamesystem: Registered > FSNamesystemStatusMBean > 2009-05-14 20:28:23,964 INFO org.apache.hadoop.dfs.Storage: Number of files > = 1 > 2009-05-14 20:28:23,971 INFO org.apache.hadoop.dfs.Storage: Number of files > under construction = 0 > 2009-05-14 20:28:23,971 INFO org.apache.hadoop.dfs.Storage: Image file of > size 80 loaded in 0 seconds. > 2009-05-14 20:28:23,972 INFO org.apache.hadoop.dfs.Storage: Edits file > edits > of size 4 edits # 0 loaded in 0 seconds. > 2009-05-14 20:28:23,974 INFO org.apache.hadoop.fs.FSNamesystem: Finished > loading FSImage in 155 msecs > 2009-05-14 20:28:23,976 INFO org.apache.hadoop.fs.FSNamesystem: Total > number > of blocks = 0 > 2009-05-14 20:28:23,988 INFO org.apache.hadoop.fs.FSNamesystem: Number of > invalid blocks = 0 > 2009-05-14 20:28:23,988 INFO org.apache.hadoop.fs.FSNamesystem: Number of > under-replicated blocks = 0 > 2009-05-14 20:28:23,988 INFO org.apache.hadoop.fs.FSNamesystem: Number of > over-replicated blocks = 0 > 2009-05-14 20:28:23,988 INFO org.apache.hadoop.dfs.StateChange: STATE* > Leaving safe mode after 0 secs. > *2009-05-14 20:28:23,989 INFO org.apache.hadoop.dfs.StateChange: STATE* > Network topology has 0 racks and 0 datanodes* > 2009-05-14 20:28:23,989 INFO org.apache.hadoop.dfs.StateChange: STATE* > UnderReplicatedBlocks has 0 blocks > 2009-05-14 20:28:29,128 INFO org.mortbay.util.Credential: Checking Resource > aliases > 2009-05-14 20:28:29,243 INFO org.mortbay.http.HttpServer: Version > Jetty/5.1.4 > 2009-05-14 20:28:29,244 INFO org.mortbay.util.Container: Started > HttpContext[/static,/static] > 2009-05-14 20:28:29,245 INFO org.mortbay.util.Container: Started > HttpContext[/logs,/logs] > 2009-05-14 20:28:29,750 INFO org.mortbay.util.Container: Started > org.mortbay.jetty.servlet.webapplicationhand...@7fcebc9f > 2009-05-14 20:28:29,838 INFO org.mortbay.util.Container: Started > WebApplicationContext[/,/] > 2009-05-14 20:28:29,843 INFO org.mortbay.http.SocketListener: Started > SocketListener on 0.0.0.0:50070 > 2009-05-14 20:28:29,843 INFO org.mortbay.util.Container: Started > org.mortbay.jetty.ser...@61acfa31 > 2009-05-14 20:28:29,843 INFO org.apache.hadoop.fs.FSNamesystem: Web-server > up at: 0.0.0.0:50070 > 2009-05-14 20:28:29,843 INFO org.apache.hadoop.ipc.Server: IPC Server > Responder: starting > 2009-05-14 20:28:29,844 INFO org.apache.hadoop.ipc.Server: IPC Server > listener on 9000: starting > 2009-05-14 20:28:29,865 INFO org.apache.hadoop.ipc.Server: IPC Server > handler 0 on 9000: starting > 2009-05-14 20:28:29,876 INFO org.apache.hadoop.ipc.Server: IPC Server > handler 1 on 9000: starting > 2009-05-14 20:28:29,877 INFO org.apache.hadoop.ipc.Server: IPC Server > handler 2 on 9000: starting > 2009-05-14 20:28:29,877 INFO org.apache.hadoop.ipc.Server: IPC Server > handler 3 on 9000: starting > 2009-05-14 20:28:29,878 INFO org.apache.hadoop.ipc.Server: IPC Server > handler 4 on 9000: starting > 2009-05-14 20:28:29,879 INFO org.apache.hadoop.ipc.Server: IPC Server > handler 5 on 9000: starting > 2009-05-14 20:28:29,879 INFO org.apache.hadoop.ipc.Server: IPC Server > handler 6 on 9000: starting > 2009-05-14 20:28:29,881 INFO org.apache.hadoop.ipc.Server: IPC Server > handler 7 on 9000: starting > 2009-05-14 20:28:29,881 INFO org.apache.hadoop.ipc.Server: IPC Server > handler 8 on 9000: starting > 2009-05-14 20:28:29,882 INFO org.apache.hadoop.ipc.Server: IPC Server > handler 9 on 9000: starting > 2009-05-14 20:33:35,774 INFO org.apache.hadoop.fs.FSNamesystem: Roll Edit > Log from 192.168.0.1 > 2009-05-14 20:33:35,775 INFO org.apache.hadoop.fs.FSNamesystem: Number of > transactions: 0 Total time for transactions(ms): 0 Number of syncs: 0 > SyncTimes(ms): 0 > 2009-05-14 20:33:36,310 INFO org.apache.hadoop.fs.FSNamesystem: Roll > FSImage > from 192.168.0.1 > 2009-05-14 20:33:36,311 INFO org.apache.hadoop.fs.FSNamesystem: Number of > transactions: 0 Total time for transactions(ms): 0 Number of syncs: 0 > SyncTimes(ms): 0 > > > > 3) my hadoop-site.xml for refrence > > <?xml version="1.0"?> > <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> > > <!-- Put site-specific property overrides in this file. --> > <configuration> > <property> > <name>fs.default.name</name> > <value>hdfs://hadoopmaster.local:9000</value> > </property> > <property> > <name>mapred.job.tracker</name> > <value>hadoopmaster.local:9001</value> > </property> > <property> > <name>dfs.replication</name> > <value>3</value> > </property> > <property> > <name>mapred.child.java.opts</name> > <value>-Xmx512m</value> > </property> > <property> > <name>hadoop.tmp.dir</name> > <value>/Hadoop/Temp</value> > <description>A base for other temporary directories.</description> > </property> > <property> > <name>dfs.data.dir</name> > <value>/Hadoop/Data,/data/Hadoop</value> > <description>Determines where on the local filesystem an DFS data node > should store its blocks. If this is a comma-delimited > list of directories, then data will be stored in all named > directories, typically on different devices. > Directories that do not exist are ignored. > </description> > </property> > </configuration> > > > the main thing i find in log it says "*2009-05-14 20:28:23,989 INFO > org.apache.hadoop.dfs.StateChange: STATE* Network topology has 0 racks and > 0 > datanodes" *which means it cannot start datanodes*.*but why is it so? I > have > all my datanodes in my slave.xml and it detects that on start up screen > when > i give start-dfs.sh. > > > Can anyone throw some lights on my problem. > > Thanks > Pankil > -- Alpha Chapters of my book on Hadoop are available http://www.apress.com/book/view/9781430219422 www.prohadoopbook.com a community for Hadoop Professionals