Hi Users, My name node is shutting down soon after it is started. Here the log. Can some one please help me.
2012-10-19 23:20:42,143 INFO org.apache.hadoop.hdfs.server.namenode.NameNode: STARTUP_MSG: /************************************************************ STARTUP_MSG: Starting NameNode STARTUP_MSG: host = sk.r252.0/10.0.2.15 STARTUP_MSG: args = [] STARTUP_MSG: version = 0.20.2 STARTUP_MSG: build = https://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20 -r 911707; compiled by 'chrisdo' on Fri Feb 19 08:07:34 UTC 2010 ************************************************************/ 2012-10-19 23:20:42,732 INFO org.apache.hadoop.ipc.metrics.RpcMetrics: Initializing RPC Metrics with hostName=NameNode, port=54310 2012-10-19 23:20:42,741 INFO org.apache.hadoop.hdfs.server.namenode.NameNode: Namenode up at: sk.r252.0/10.0.2.15:54310 2012-10-19 23:20:42,745 INFO org.apache.hadoop.metrics.jvm.JvmMetrics: Initializing JVM Metrics with processName=NameNode, sessionId=null 2012-10-19 23:20:42,747 INFO org.apache.hadoop.hdfs.server.namenode.metrics.NameNodeMetrics: Initializing NameNodeMeterics using context object:org.apache.hadoop.metrics.spi.NullContext 2012-10-19 23:20:43,074 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: fsOwner=root,root,bin,daemon,sys,adm,disk,wheel 2012-10-19 23:20:43,077 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: supergroup=supergroup 2012-10-19 23:20:43,077 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: isPermissionEnabled=true 2012-10-19 23:20:43,231 INFO org.apache.hadoop.hdfs.server.namenode.metrics.FSNamesystemMetrics: Initializing FSNamesystemMetrics using context object:org.apache.hadoop.metrics.spi.NullContext 2012-10-19 23:20:43,239 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Registered FSNamesystemStatusMBean 2012-10-19 23:20:43,359 INFO org.apache.hadoop.hdfs.server.common.Storage: Number of files = 1 2012-10-19 23:20:43,379 INFO org.apache.hadoop.hdfs.server.common.Storage: Number of files under construction = 0 2012-10-19 23:20:43,379 INFO org.apache.hadoop.hdfs.server.common.Storage: Image file of size 94 loaded in 0 seconds. 2012-10-19 23:20:43,380 INFO org.apache.hadoop.hdfs.server.common.Storage: Edits file /app/hadoop/tmp/dfs/name/current/edits of size 4 edits # 0 loaded in 0 seconds. 2012-10-19 23:20:43,415 INFO org.apache.hadoop.hdfs.server.common.Storage: Image file of size 94 saved in 0 seconds. 2012-10-19 23:20:43,612 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Finished loading FSImage in 758 msecs 2012-10-19 23:20:43,615 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Total number of blocks = 0 2012-10-19 23:20:43,615 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of invalid blocks = 0 2012-10-19 23:20:43,615 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of under-replicated blocks = 0 2012-10-19 23:20:43,615 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of over-replicated blocks = 0 2012-10-19 23:20:43,615 INFO org.apache.hadoop.hdfs.StateChange: STATE* Leaving safe mode after 0 secs. 2012-10-19 23:20:43,616 INFO org.apache.hadoop.hdfs.StateChange: STATE* Network topology has 0 racks and 0 datanodes 2012-10-19 23:20:43,616 INFO org.apache.hadoop.hdfs.StateChange: STATE* UnderReplicatedBlocks has 0 blocks 2012-10-19 23:20:44,450 INFO org.mortbay.log: Logging to org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via org.mortbay.log.Slf4jLog 2012-10-19 23:20:44,711 INFO org.apache.hadoop.http.HttpServer: Port returned by webServer.getConnectors()[0].getLocalPort() before open() is -1. Opening the listener on 50070 2012-10-19 23:20:44,715 INFO org.apache.hadoop.http.HttpServer: listener.getLocalPort() returned 50070 webServer.getConnectors()[0].getLocalPort() returned 50070 2012-10-19 23:20:44,715 INFO org.apache.hadoop.http.HttpServer: Jetty bound to port 50070 2012-10-19 23:20:44,715 INFO org.mortbay.log: jetty-6.1.14 2012-10-19 23:20:47,021 INFO org.mortbay.log: Started [email protected]:50070 2012-10-19 23:20:47,022 INFO org.apache.hadoop.hdfs.server.namenode.NameNode: Web-server up at: 0.0.0.0:50070 2012-10-19 23:20:47,022 INFO org.apache.hadoop.hdfs.server.namenode.NameNode: Web-server up at: 0.0.0.0:50070 2012-10-19 23:20:47,067 INFO org.apache.hadoop.ipc.Server: IPC Server listener on 54310: starting 2012-10-19 23:20:47,086 INFO org.apache.hadoop.ipc.Server: IPC Server handler 0 on 54310: starting 2012-10-19 23:20:47,089 INFO org.apache.hadoop.ipc.Server: IPC Server Responder: starting 2012-10-19 23:20:47,106 INFO org.apache.hadoop.ipc.Server: IPC Server handler 1 on 54310: starting 2012-10-19 23:20:47,130 INFO org.apache.hadoop.ipc.Server: IPC Server handler 2 on 54310: starting 2012-10-19 23:20:47,148 INFO org.apache.hadoop.ipc.Server: IPC Server handler 3 on 54310: starting 2012-10-19 23:20:47,165 INFO org.apache.hadoop.ipc.Server: IPC Server handler 4 on 54310: starting 2012-10-19 23:20:47,183 INFO org.apache.hadoop.ipc.Server: IPC Server handler 5 on 54310: starting 2012-10-19 23:20:47,200 INFO org.apache.hadoop.ipc.Server: IPC Server handler 6 on 54310: starting 2012-10-19 23:20:47,803 INFO org.apache.hadoop.ipc.Server: IPC Server handler 9 on 54310: starting 2012-10-19 23:20:47,804 INFO org.apache.hadoop.ipc.Server: IPC Server handler 7 on 54310: starting 2012-10-19 23:20:47,806 INFO org.apache.hadoop.ipc.Server: IPC Server handler 8 on 54310: starting 2012-10-19 23:20:48,685WARN org.apache.hadoop.hdfs.server.namenode.FSNamesystem: ReplicationMonitor thread received InterruptedException.java.lang.InterruptedException: sleep interrupted 2012-10-19 23:20:48,691 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of transactions: 0 Total time for transactions(ms): 0Number of transactions batched in Syncs: 0 Number of syncs: 0 SyncTimes(ms): 0 2012-10-19 23:20:48,690 INFO org.apache.hadoop.hdfs.server.namenode.DecommissionManager: Interrupted Monitor java.lang.InterruptedException: sleep interrupted at java.lang.Thread.sleep(Native Method) at org.apache.hadoop.hdfs.server.namenode.DecommissionManager$Monitor.run(DecommissionManager.java:65) at java.lang.Thread.run(Thread.java:636) 2012-10-19 23:20:48,771 INFO org.apache.hadoop.ipc.Server: Stopping server on 54310 2012-10-19 23:20:48,775 INFO org.apache.hadoop.ipc.Server: IPC Server handler 0 on 54310: exiting 2012-10-19 23:20:48,780 INFO org.apache.hadoop.ipc.Server: IPC Server handler 1 on 54310: exiting 2012-10-19 23:20:48,781 INFO org.apache.hadoop.ipc.Server: IPC Server handler 2 on 54310: exiting 2012-10-19 23:20:48,782 INFO org.apache.hadoop.ipc.Server: IPC Server handler 3 on 54310: exiting 2012-10-19 23:20:48,783 INFO org.apache.hadoop.ipc.Server: IPC Server handler 4 on 54310: exiting 2012-10-19 23:20:48,784 INFO org.apache.hadoop.ipc.Server: IPC Server handler 5 on 54310: exiting 2012-10-19 23:20:48,784 INFO org.apache.hadoop.ipc.Server: IPC Server handler 6 on 54310: exiting 2012-10-19 23:20:48,785 INFO org.apache.hadoop.ipc.Server: IPC Server handler 7 on 54310: exiting 2012-10-19 23:20:48,785 INFO org.apache.hadoop.ipc.Server: IPC Server handler 8 on 54310: exiting 2012-10-19 23:20:48,786 INFO org.apache.hadoop.ipc.Server: IPC Server handler 9 on 54310: exiting 2012-10-19 23:20:48,786 INFO org.apache.hadoop.ipc.Server: Stopping IPC Server listener on 54310 2012-10-19 23:20:48,788 INFO org.apache.hadoop.ipc.Server: Stopping IPC Server Responder 2012-10-19 23:20:48,790 ERROR org.apache.hadoop.hdfs.server.namenode.NameNode: java.io.IOException: Incomplete HDFS URI, no host: hdfs://sk.r252.0:54310 at org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:78) at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1378) at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:66) at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1390) at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:196) at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:95) at org.apache.hadoop.fs.Trash.<init>(Trash.java:62) at org.apache.hadoop.hdfs.server.namenode.NameNode.startTrashEmptier(NameNode.java:208) at org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:204) at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:279) at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:956) at org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:965) 2012-10-19 23:20:48,995 INFO org.apache.hadoop.hdfs.server.namenode.NameNode: SHUTDOWN_MSG: /************************************************************ SHUTDOWN_MSG: Shutting down NameNode at sk.r252.0/10.0.2.15 ***core-site.xml*** <?xml version="1.0"?> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> <!-- Put site-specific property overrides in this file. --> <configuration> <property> <name>hadoop.tmp.dir</name> <value>/app/hadoop/tmp</value> <description>A base for other temporary directories.</description> </property> <property> <name>fs.default.name</name> <value>hdfs://sk.r252.0:54310</value> <description>The name of the default file system. A URI whose scheme and authority determine the FileSystem implementation. The uri's scheme determines the config property (fs.SCHEME.impl) naming the FileSystem implementation class. The uri's authority is used to determine the host, port, etc. for a filesystem.</description> </property> </configuration> ***mapred-site.xml*** <?xml version="1.0"?> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> <!-- Put site-specific property overrides in this file. --> <configuration> <property> <name>mapred.job.tracker</name> <value>sk.r252.0:54311</value> <description>The host and port that the MapReduce job tracker runs at. If "local", then jobs are run in-process as a single map and reduce task. </description> </property> </configuration> ***hdfs-site.xml*** <?xml version="1.0"?> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> <!-- Put site-specific property overrides in this file. --> <configuration> <property> <name>dfs.replication</name> <value>2</value> <description>Default block replication. The actual number of replications can be specified when the file is created. The default is used if replication is not specified in create time. </description> </property> <property> <name>dfs.http.address</name> <value>0.0.0.0:50070</value> </property> </configuration> Can some one please help me. Regards Sundeep
