Thank  You Balaji,
I checked gethostbyname(sk.r252.0) it gives 10.0.2.15. This is ipaddress i am 
getting in ifconfig also.
ssh sk.r252.0 is sshing to 10.0.2.15
ping sk.r252.0 is pinging to 10.0.2.15.

Can you please help me with the issue?

Regards
Sundeep



________________________________
 From: Balaji Narayanan (பாலாஜி நாராயணன்) <[email protected]>
To: "[email protected]" <[email protected]>; Sundeep Kambhmapati 
<[email protected]> 
Sent: Saturday, 20 October 2012 2:12 AM
Subject: Re: Namenode shutting down while creating cluster
 

Seems like an issue with resolution of sk.r252.0. Can you ensure that it 
resolves?

On Friday, October 19, 2012, Sundeep Kambhmapati  wrote:

Hi Users,
>My name node is shutting down soon after it is started. 
>Here the log. Can some one please help me.
>
>
>2012-10-19 23:20:42,143 INFO org.apache.hadoop.hdfs.server.namenode.NameNode: 
>STARTUP_MSG:
>/************************************************************
>STARTUP_MSG: Starting NameNode
>STARTUP_MSG:   host = sk.r252.0/10.0.2.15
>STARTUP_MSG:   args = []
>STARTUP_MSG:   version = 0.20.2
>STARTUP_MSG:   build = 
>https://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20 -r 911707; 
>compiled by 'chrisdo' on Fri Feb 19 08:07:34 UTC 2010
>************************************************************/
>2012-10-19 23:20:42,732 INFO org.apache.hadoop.ipc.metrics.RpcMetrics: 
>Initializing RPC Metrics with hostName=NameNode, port=54310
>2012-10-19 23:20:42,741 INFO org.apache.hadoop.hdfs.server.namenode.NameNode: 
>Namenode up at: sk.r252.0/10.0.2.15:54310
>2012-10-19 23:20:42,745 INFO org.apache.hadoop.metrics.jvm.JvmMetrics: 
>Initializing JVM Metrics with processName=NameNode, sessionId=null
>2012-10-19 23:20:42,747 INFO 
>org.apache.hadoop.hdfs.server.namenode.metrics.NameNodeMetrics: Initializing 
>NameNodeMeterics using context object:org.apache.hadoop.metrics.spi.NullContext
>2012-10-19 23:20:43,074 INFO 
>org.apache.hadoop.hdfs.server.namenode.FSNamesystem: 
>fsOwner=root,root,bin,daemon,sys,adm,disk,wheel
>2012-10-19 23:20:43,077 INFO 
>org.apache.hadoop.hdfs.server.namenode.FSNamesystem: supergroup=supergroup
>2012-10-19 23:20:43,077 INFO 
>org.apache.hadoop.hdfs.server.namenode.FSNamesystem: isPermissionEnabled=true
>2012-10-19 23:20:43,231 INFO 
>org.apache.hadoop.hdfs.server.namenode.metrics.FSNamesystemMetrics: 
>Initializing FSNamesystemMetrics using context 
>object:org.apache.hadoop.metrics.spi.NullContext
>2012-10-19 23:20:43,239 INFO 
>org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Registered 
>FSNamesystemStatusMBean
>2012-10-19 23:20:43,359 INFO org.apache.hadoop.hdfs.server.common.Storage: 
>Number of files = 1
>2012-10-19 23:20:43,379 INFO org.apache.hadoop.hdfs.server.common.Storage: 
>Number of files under construction = 0
>2012-10-19 23:20:43,379 INFO org.apache.hadoop.hdfs.server.common.Storage: 
>Image file of size 94 loaded in 0 seconds.
>2012-10-19 23:20:43,380 INFO org.apache.hadoop.hdfs.server.common.Storage: 
>Edits file /app/hadoop/tmp/dfs/name/current/edits of size 4 edits # 0 loaded 
>in 0 seconds.
>2012-10-19 23:20:43,415 INFO org.apache.hadoop.hdfs.server.common.Storage: 
>Image file of size 94 saved in 0 seconds.
>2012-10-19 23:20:43,612 INFO 
>org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Finished loading FSImage 
>in 758 msecs
>2012-10-19 23:20:43,615 INFO 
>org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Total number of blocks = 0
>2012-10-19 23:20:43,615 INFO 
>org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of invalid blocks 
>= 0
>2012-10-19 23:20:43,615 INFO 
>org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of 
>under-replicated blocks = 0
>2012-10-19 23:20:43,615 INFO 
>org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of  
>over-replicated blocks = 0
>2012-10-19 23:20:43,615 INFO org.apache.hadoop.hdfs.StateChange: STATE* 
>Leaving safe mode after 0 secs.
>2012-10-19 23:20:43,616 INFO org.apache.hadoop.hdfs.StateChange: STATE* 
>Network topology has 0 racks and 0 datanodes
>2012-10-19 23:20:43,616 INFO org.apache.hadoop.hdfs.StateChange: STATE* 
>UnderReplicatedBlocks has 0 blocks
>2012-10-19 23:20:44,450 INFO org.mortbay.log: Logging to 
>org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via org.mortbay.log.Slf4jLog
>2012-10-19 23:20:44,711 INFO org.apache.hadoop.http.HttpServer: Port returned 
>by webServer.getConnectors()[0].getLocalPort() before open() is -1. Opening 
>the listener on 50070
>2012-10-19 23:20:44,715 INFO org.apache.hadoop.http.HttpServer: 
>listener.getLocalPort() returned 50070 
>webServer.getConnectors()[0].getLocalPort() returned 50070
>2012-10-19 23:20:44,715 INFO org.apache.hadoop.http.HttpServer: Jetty bound to 
>port 50070
>2012-10-19 23:20:44,715 INFO org.mortbay.log: jetty-6.1.14
>2012-10-19 23:20:47,021 INFO org.mortbay.log: Started 
>[email protected]:50070
>2012-10-19 23:20:47,022 INFO org.apache.hadoop.hdfs.server.namenode.NameNode: 
>Web-server up at: 0.0.0.0:50070
>2012-10-19 23:20:47,022 INFO org.apache.hadoop.hdfs.server.namenode.NameNode: 
>Web-server up at: 0.0.0.0:50070
>2012-10-19 23:20:47,067 INFO org.apache.hadoop.ipc.Server: IPC Server listener 
>on 54310: starting
>2012-10-19 23:20:47,086 INFO org.apache.hadoop.ipc.Server: IPC Server handler 
>0 on 54310: starting
>2012-10-19 23:20:47,089 INFO org.apache.hadoop.ipc.Server: IPC Server 
>Responder: starting
>2012-10-19 23:20:47,106 INFO org.apache.hadoop.ipc.Server: IPC Server handler 
>1 on 54310: starting
>2012-10-19 23:20:47,130 INFO org.apache.hadoop.ipc.Server: IPC Server handler 
>2 on 54310: starting
>2012-10-19 23:20:47,148 INFO org.apache.hadoop.ipc.Server: IPC Server handler 
>3 on 54310: starting
>2012-10-19 23:20:47,165 INFO org.apache.hadoop.ipc.Server: IPC Server handler 
>4 on 54310: starting
>2012-10-19 23:20:47,183 INFO org.apache.hadoop.ipc.Server: IPC Server handler 
>5 on 54310: starting
>2012-10-19 23:20:47,200 INFO org.apache.hadoop.ipc.Server: IPC Server handler 
>6 on 54310: starting
>2012-10-19 23:20:47,803 INFO org.apache.hadoop.ipc.Server: IPC Server handler 
>9 on 54310: starting
>2012-10-19 23:20:47,804 INFO org.apache.hadoop.ipc.Server: IPC Server handler 
>7 on 54310: starting
>2012-10-19 23:20:47,806 INFO org.apache.hadoop.ipc.Server: IPC Server handler 
>8 on 54310: starting
>2012-10-19 23:20:48,685WARN 
>org.apache.hadoop.hdfs.server.namenode.FSNamesystem: ReplicationMonitor thread 
>received InterruptedException.java.lang.InterruptedException: sleep interrupted
>2012-10-19 23:20:48,691 INFO 
>org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of transactions: 0 
>Total time for transactions(ms): 0Number of transactions batched in Syncs: 0 
>Number of syncs: 0 SyncTimes(ms): 0
>2012-10-19 23:20:48,690 INFO 
>org.apache.hadoop.hdfs.server.namenode.DecommissionManager: Interrupted Monitor
>java.lang.InterruptedException: sleep interrupted
>        at java.lang.Thread.sleep(Native Method)
>        at 
>org.apache.hadoop.hdfs.server.namenode.DecommissionManager$Monitor.run(DecommissionManager.java:65)
>        at java.lang.Thread.run(Thread.java:636)
>2012-10-19 23:20:48,771 INFO org.apache.hadoop.ipc.Server: Stopping server on 
>54310
>2012-10-19 23:20:48,775 INFO org.apache.hadoop.ipc.Server: IPC Server handler 
>0 on 54310: exiting
>2012-10-19 23:20:48,780 INFO org.apache.hadoop.ipc.Server: IPC Server handler 
>1 on 54310: exiting
>2012-10-19 23:20:48,781 INFO org.apache.hadoop.ipc.Server: IPC Server handler 
>2 on 54310: exiting
>2012-10-19 23:20:48,782 INFO org.apache.hadoop.ipc.Server: IPC Server handler 
>3 on 54310: exiting
>2012-10-19 23:20:48,783 INFO org.apache.hadoop.ipc.Server: IPC Server handler 
>4 on 54310: exiting
>2012-10-19 23:20:48,784 INFO org.apache.hadoop.ipc.Server: IPC Server handler 
>5 on 54310: exiting
>2012-10-19 23:20:48,784 INFO org.apache.hadoop.ipc.Server: IPC Server handler 
>6 on 54310: exiting
>2012-10-19 23:20:48,785 INFO org.apache.hadoop.ipc.Server: IPC Server handler 
>7 on 54310: exiting
>2012-10-19 23:20:48,785 INFO org.apache.hadoop.ipc.Server: IPC Server handler 
>8 on 54310: exiting
>2012-10-19 23:20:48,786 INFO org.apache.hadoop.ipc.Server: IPC Server handler 
>9 on 54310: exiting
>2012-10-19 23:20:48,786 INFO org.apache.hadoop.ipc.Server: Stopping IPC Server 
>listener on 54310
>2012-10-19 23:20:48,788 INFO org.apache.hadoop.ipc.Server: Stopping IPC Server 
>Responder
>2012-10-19 23:20:48,790 ERROR org.apache.hadoop.hdfs.server.namenode.NameNode: 
>java.io.IOException: Incomplete HDFS URI, no host: hdfs://sk.r252.0:54310
>        at 
>org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:78)
>        at 
>org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1378)
>        at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:66)
>        at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1390)
>        at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:196)
>        at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:95)
>         at org.apache.hadoop.fs.Trash.<init>(Trash.java:62)
>        at 
>org.apache.hadoop.hdfs.server.namenode.NameNode.startTrashEmptier(NameNode.java:208)
>        at 
>org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:204)
>        at 
>org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:279)
>        at 
>org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:956)
>        at 
>org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:965)
>
>
>2012-10-19 23:20:48,995 INFO org.apache.hadoop.hdfs.server.namenode.NameNode: 
>SHUTDOWN_MSG:
>/************************************************************
>SHUTDOWN_MSG: Shutting down NameNode at sk.r252.0/10.0.2.15
>
>
>***core-site.xml***
><?xml version="1.0"?>
><?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
>
>
><!-- Put site-specific property overrides in this file. -->
>
>
><configuration>
><property>
>  <name>hadoop.tmp.dir</name>
>  <value>/app/hadoop/tmp</value>
>  <description>A base for other temporary directories.</description>
></property>
>
>
><property>
>  <name>fs.default.name</name>
>  <value>hdfs://sk.r252.0:54310</value>
>  <description>The name of the default file system.  A URI whose
>  scheme and authority determine the FileSystem implementation.  The
>  uri's scheme determines the config property (fs.SCHEME.impl) naming
>  the FileSystem implementation class.  The uri's authority is used to
>  determine the host, port, etc. for a filesystem.</description>
></property>
></configuration>
>
>
>***mapred-site.xml***
><?xml version="1.0"?>
><?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
>
>
><!-- Put site-specific property overrides in this file. -->
>
>
><configuration>
><property>
>  <name>mapred.job.tracker</name>
>  <value>sk.r252.0:54311</value>
>  <description>The host and port that the MapReduce job tracker runs
>  at.  If "local", then jobs are run in-process as a single map
>  and reduce task.
>  </description>
></property>
></configuration>
>
>
>***hdfs-site.xml***
><?xml version="1.0"?>
><?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
>
>
><!-- Put site-specific property overrides in this file. -->
>
>
><configuration>
><property>
>  <name>dfs.replication</name>
>  <value>2</value>
>  <description>Default block replication.
>  The actual number of replications can be specified when the file is created.
>  The default is used if replication is not specified in create time.
>  </description>
></property>
><property>
>    <name>dfs.http.address</name>
>    <value>0.0.0.0:50070</value>
>  </property>
></configuration>
>
>
>Can some one please help me.
>
>
>Regards 
>Sundeep
>
>

-- 
Thanks
-balaji
--
http://balajin.net/blog/
http://flic.kr/balajijegan

Reply via email to