Hi Users,
My name node is shutting down soon after it is started. 
Here the log. Can some one please help me.

2012-10-19 23:20:42,143 INFO org.apache.hadoop.hdfs.server.namenode.NameNode: 
STARTUP_MSG:
/************************************************************
STARTUP_MSG: Starting NameNode
STARTUP_MSG:   host = sk.r252.0/10.0.2.15
STARTUP_MSG:   args = []
STARTUP_MSG:   version = 0.20.2
STARTUP_MSG:   build = 
https://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20 -r 911707; 
compiled by 'chrisdo' on Fri Feb 19 08:07:34 UTC 2010
************************************************************/
2012-10-19 23:20:42,732 INFO org.apache.hadoop.ipc.metrics.RpcMetrics: 
Initializing RPC Metrics with hostName=NameNode, port=54310
2012-10-19 23:20:42,741 INFO org.apache.hadoop.hdfs.server.namenode.NameNode: 
Namenode up at: sk.r252.0/10.0.2.15:54310
2012-10-19 23:20:42,745 INFO org.apache.hadoop.metrics.jvm.JvmMetrics: 
Initializing JVM Metrics with processName=NameNode, sessionId=null
2012-10-19 23:20:42,747 INFO 
org.apache.hadoop.hdfs.server.namenode.metrics.NameNodeMetrics: Initializing 
NameNodeMeterics using context object:org.apache.hadoop.metrics.spi.NullContext
2012-10-19 23:20:43,074 INFO 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem: 
fsOwner=root,root,bin,daemon,sys,adm,disk,wheel
2012-10-19 23:20:43,077 INFO 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem: supergroup=supergroup
2012-10-19 23:20:43,077 INFO 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem: isPermissionEnabled=true
2012-10-19 23:20:43,231 INFO 
org.apache.hadoop.hdfs.server.namenode.metrics.FSNamesystemMetrics: 
Initializing FSNamesystemMetrics using context 
object:org.apache.hadoop.metrics.spi.NullContext
2012-10-19 23:20:43,239 INFO 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Registered 
FSNamesystemStatusMBean
2012-10-19 23:20:43,359 INFO org.apache.hadoop.hdfs.server.common.Storage: 
Number of files = 1
2012-10-19 23:20:43,379 INFO org.apache.hadoop.hdfs.server.common.Storage: 
Number of files under construction = 0
2012-10-19 23:20:43,379 INFO org.apache.hadoop.hdfs.server.common.Storage: 
Image file of size 94 loaded in 0 seconds.
2012-10-19 23:20:43,380 INFO org.apache.hadoop.hdfs.server.common.Storage: 
Edits file /app/hadoop/tmp/dfs/name/current/edits of size 4 edits # 0 loaded in 
0 seconds.
2012-10-19 23:20:43,415 INFO org.apache.hadoop.hdfs.server.common.Storage: 
Image file of size 94 saved in 0 seconds.
2012-10-19 23:20:43,612 INFO 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Finished loading FSImage 
in 758 msecs
2012-10-19 23:20:43,615 INFO 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Total number of blocks = 0
2012-10-19 23:20:43,615 INFO 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of invalid blocks = 0
2012-10-19 23:20:43,615 INFO 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of under-replicated 
blocks = 0
2012-10-19 23:20:43,615 INFO 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of  over-replicated 
blocks = 0
2012-10-19 23:20:43,615 INFO org.apache.hadoop.hdfs.StateChange: STATE* Leaving 
safe mode after 0 secs.
2012-10-19 23:20:43,616 INFO org.apache.hadoop.hdfs.StateChange: STATE* Network 
topology has 0 racks and 0 datanodes
2012-10-19 23:20:43,616 INFO org.apache.hadoop.hdfs.StateChange: STATE* 
UnderReplicatedBlocks has 0 blocks
2012-10-19 23:20:44,450 INFO org.mortbay.log: Logging to 
org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via org.mortbay.log.Slf4jLog
2012-10-19 23:20:44,711 INFO org.apache.hadoop.http.HttpServer: Port returned 
by webServer.getConnectors()[0].getLocalPort() before open() is -1. Opening the 
listener on 50070
2012-10-19 23:20:44,715 INFO org.apache.hadoop.http.HttpServer: 
listener.getLocalPort() returned 50070 
webServer.getConnectors()[0].getLocalPort() returned 50070
2012-10-19 23:20:44,715 INFO org.apache.hadoop.http.HttpServer: Jetty bound to 
port 50070
2012-10-19 23:20:44,715 INFO org.mortbay.log: jetty-6.1.14
2012-10-19 23:20:47,021 INFO org.mortbay.log: Started 
[email protected]:50070
2012-10-19 23:20:47,022 INFO org.apache.hadoop.hdfs.server.namenode.NameNode: 
Web-server up at: 0.0.0.0:50070
2012-10-19 23:20:47,022 INFO org.apache.hadoop.hdfs.server.namenode.NameNode: 
Web-server up at: 0.0.0.0:50070
2012-10-19 23:20:47,067 INFO org.apache.hadoop.ipc.Server: IPC Server listener 
on 54310: starting
2012-10-19 23:20:47,086 INFO org.apache.hadoop.ipc.Server: IPC Server handler 0 
on 54310: starting
2012-10-19 23:20:47,089 INFO org.apache.hadoop.ipc.Server: IPC Server 
Responder: starting
2012-10-19 23:20:47,106 INFO org.apache.hadoop.ipc.Server: IPC Server handler 1 
on 54310: starting
2012-10-19 23:20:47,130 INFO org.apache.hadoop.ipc.Server: IPC Server handler 2 
on 54310: starting
2012-10-19 23:20:47,148 INFO org.apache.hadoop.ipc.Server: IPC Server handler 3 
on 54310: starting
2012-10-19 23:20:47,165 INFO org.apache.hadoop.ipc.Server: IPC Server handler 4 
on 54310: starting
2012-10-19 23:20:47,183 INFO org.apache.hadoop.ipc.Server: IPC Server handler 5 
on 54310: starting
2012-10-19 23:20:47,200 INFO org.apache.hadoop.ipc.Server: IPC Server handler 6 
on 54310: starting
2012-10-19 23:20:47,803 INFO org.apache.hadoop.ipc.Server: IPC Server handler 9 
on 54310: starting
2012-10-19 23:20:47,804 INFO org.apache.hadoop.ipc.Server: IPC Server handler 7 
on 54310: starting
2012-10-19 23:20:47,806 INFO org.apache.hadoop.ipc.Server: IPC Server handler 8 
on 54310: starting
2012-10-19 23:20:48,685WARN 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem: ReplicationMonitor thread 
received InterruptedException.java.lang.InterruptedException: sleep interrupted
2012-10-19 23:20:48,691 INFO 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of transactions: 0 
Total time for transactions(ms): 0Number of transactions batched in Syncs: 0 
Number of syncs: 0 SyncTimes(ms): 0
2012-10-19 23:20:48,690 INFO 
org.apache.hadoop.hdfs.server.namenode.DecommissionManager: Interrupted Monitor
java.lang.InterruptedException: sleep interrupted
        at java.lang.Thread.sleep(Native Method)
        at 
org.apache.hadoop.hdfs.server.namenode.DecommissionManager$Monitor.run(DecommissionManager.java:65)
        at java.lang.Thread.run(Thread.java:636)
2012-10-19 23:20:48,771 INFO org.apache.hadoop.ipc.Server: Stopping server on 
54310
2012-10-19 23:20:48,775 INFO org.apache.hadoop.ipc.Server: IPC Server handler 0 
on 54310: exiting
2012-10-19 23:20:48,780 INFO org.apache.hadoop.ipc.Server: IPC Server handler 1 
on 54310: exiting
2012-10-19 23:20:48,781 INFO org.apache.hadoop.ipc.Server: IPC Server handler 2 
on 54310: exiting
2012-10-19 23:20:48,782 INFO org.apache.hadoop.ipc.Server: IPC Server handler 3 
on 54310: exiting
2012-10-19 23:20:48,783 INFO org.apache.hadoop.ipc.Server: IPC Server handler 4 
on 54310: exiting
2012-10-19 23:20:48,784 INFO org.apache.hadoop.ipc.Server: IPC Server handler 5 
on 54310: exiting
2012-10-19 23:20:48,784 INFO org.apache.hadoop.ipc.Server: IPC Server handler 6 
on 54310: exiting
2012-10-19 23:20:48,785 INFO org.apache.hadoop.ipc.Server: IPC Server handler 7 
on 54310: exiting
2012-10-19 23:20:48,785 INFO org.apache.hadoop.ipc.Server: IPC Server handler 8 
on 54310: exiting
2012-10-19 23:20:48,786 INFO org.apache.hadoop.ipc.Server: IPC Server handler 9 
on 54310: exiting
2012-10-19 23:20:48,786 INFO org.apache.hadoop.ipc.Server: Stopping IPC Server 
listener on 54310
2012-10-19 23:20:48,788 INFO org.apache.hadoop.ipc.Server: Stopping IPC Server 
Responder
2012-10-19 23:20:48,790 ERROR org.apache.hadoop.hdfs.server.namenode.NameNode: 
java.io.IOException: Incomplete HDFS URI, no host: hdfs://sk.r252.0:54310
        at 
org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:78)
        at 
org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1378)
        at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:66)
        at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1390)
        at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:196)
        at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:95)
         at org.apache.hadoop.fs.Trash.<init>(Trash.java:62)
        at 
org.apache.hadoop.hdfs.server.namenode.NameNode.startTrashEmptier(NameNode.java:208)
        at 
org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:204)
        at 
org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:279)
        at 
org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:956)
        at 
org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:965)

2012-10-19 23:20:48,995 INFO org.apache.hadoop.hdfs.server.namenode.NameNode: 
SHUTDOWN_MSG:
/************************************************************
SHUTDOWN_MSG: Shutting down NameNode at sk.r252.0/10.0.2.15

***core-site.xml***
<?xml version="1.0"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>

<!-- Put site-specific property overrides in this file. -->

<configuration>
<property>
  <name>hadoop.tmp.dir</name>
  <value>/app/hadoop/tmp</value>
  <description>A base for other temporary directories.</description>
</property>

<property>
  <name>fs.default.name</name>
  <value>hdfs://sk.r252.0:54310</value>
  <description>The name of the default file system.  A URI whose
  scheme and authority determine the FileSystem implementation.  The
  uri's scheme determines the config property (fs.SCHEME.impl) naming
  the FileSystem implementation class.  The uri's authority is used to
  determine the host, port, etc. for a filesystem.</description>
</property>
</configuration>

***mapred-site.xml***
<?xml version="1.0"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>

<!-- Put site-specific property overrides in this file. -->

<configuration>
<property>
  <name>mapred.job.tracker</name>
  <value>sk.r252.0:54311</value>
  <description>The host and port that the MapReduce job tracker runs
  at.  If "local", then jobs are run in-process as a single map
  and reduce task.
  </description>
</property>
</configuration>

***hdfs-site.xml***
<?xml version="1.0"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>

<!-- Put site-specific property overrides in this file. -->

<configuration>
<property>
  <name>dfs.replication</name>
  <value>2</value>
  <description>Default block replication.
  The actual number of replications can be specified when the file is created.
  The default is used if replication is not specified in create time.
  </description>
</property>
<property>
    <name>dfs.http.address</name>
    <value>0.0.0.0:50070</value>
  </property>
</configuration>

Can some one please help me.

Regards 
Sundeep

Reply via email to