Have you done the following? bin/hadoop namenode -format
On Thu, Aug 11, 2011 at 10:50 AM, A Df <[email protected]> wrote: > Hello Again: > I extracted hadoop and changed the xml as shwon in the tutorial but now it > seems it connect get a connection. I am using putty to ssh to the server and > I change the config files to set it up in pseudo mode as shown > conf/core-site.xml: > <configuration> > <property> > <name>fs.default.name</name> > <value>hdfs://localhost:9000</ > value> > </property> > </configuration> > hdfs-site.xml: > <configuration> > <property> > <name>dfs.replication</name> > <value>1</value> > </property> > <property> > <name>dfs.http.address</name> > <value>0.0.0.0:3500</value> > </property> > </configuration> > > conf/mapred-site.xml: > <configuration> > <property> > <name>mapred.job.tracker</name> > <value>localhost:9001</value> > </property> > <property> > <name>mapred.job.tracker.http.address</name> > <value>0.0.0.0:3501</value> > </property> > </configuration> > > I tried to format the namenode, started all processes but I notice that when > I stop it, it said that the namenode was not running.When I try to run the > example jar, it keeps timing out when connecting to 127.0.0.1:port#. I used > various port numbers and tried replacing localhost with the name for the > server but it still times out. It also has a long ip address for > name.server.ac.uk/161.74.12.97:3000 which seems to repeating itself since > name.server.ac.uk already has the ip address of 161.74.12.97. The console > message is shown below. I was also having problems where it did not want to > format the namenode. > > Is there something is wrong with connecting to the namenode and what cause > it to not format? > > > 2011-08-11 05:49:13,529 INFO > org.apache.hadoop.hdfs.server.namenode.NameNode: STARTUP_MSG: > /************************************************************ > STARTUP_MSG: Starting NameNode > STARTUP_MSG: host = name.server.ac.uk/161.74.12.97 > STARTUP_MSG: args = [] > STARTUP_MSG: version = 0.20.2 > STARTUP_MSG: build = > https://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20 -r > 911707; compiled by 'chrisdo' on Fri Feb 19 08:07:34 UTC 2010 > ************************************************************/ > 2011-08-11 05:49:13,663 INFO org.apache.hadoop.ipc.metrics.RpcMetrics: > Initializing RPC Metrics with hostName=NameNode, port=3000 > 2011-08-11 05:49:13,669 INFO > org.apache.hadoop.hdfs.server.namenode.NameNode: Namenode up at: > name.server.ac.uk/161.74.12.97:3000 > 2011-08-11 05:49:13,672 INFO org.apache.hadoop.metrics.jvm.JvmMetrics: > Initializing JVM Metrics with processName=NameNode, sessionId=null > 2011-08-11 05:49:13,674 INFO > org.apache.hadoop.hdfs.server.namenode.metrics.NameNodeMetrics: Initializing > NameNodeMeterics using context > object:org.apache.hadoop.metrics.spi.NullContext > 2011-08-11 05:49:13,755 INFO > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: > fsOwner=w1153435,users,cluster_login > 2011-08-11 05:49:13,755 INFO > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: supergroup=supergroup > 2011-08-11 05:49:13,756 INFO > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: > isPermissionEnabled=true > 2011-08-11 05:49:13,768 INFO > org.apache.hadoop.hdfs.server.namenode.metrics.FSNamesystemMetrics: > Initializing FSNamesystemMetrics using context > object:org.apache.hadoop.metrics.spi.NullContext > 2011-08-11 05:49:13,770 INFO > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Registered > FSNamesystemStatusMBean > 2011-08-11 05:49:13,812 ERROR > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: FSNamesystem > initialization failed. > java.io.IOException: NameNode is not formatted. > at > org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:317) > at > org.apache.hadoop.hdfs.server.namenode.FSDirectory.loadFSImage(FSDirectory.java:87) > at > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.initialize(FSNamesystem.java:311) > at > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.<init>(FSNamesystem.java:292) > at > org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:201) > at > org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:279) > at > org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:956) > at > org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:965) > 2011-08-11 05:49:13,813 INFO org.apache.hadoop.ipc.Server: Stopping server > on 3000 > 2011-08-11 05:49:13,814 ERROR > org.apache.hadoop.hdfs.server.namenode.NameNode: java.io.IOException: > NameNode is not formatted. > at > org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:317) > at > org.apache.hadoop.hdfs.server.namenode.FSDirectory.loadFSImage(FSDirectory.java:87) > at > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.initialize(FSNamesystem.java:311) > at > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.<init>(FSNamesystem.java:292) > at > org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:201) > at > org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:279) > at > org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:956) > at > org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:965) > > 2011-08-11 05:49:13,814 INFO > org.apache.hadoop.hdfs.server.namenode.NameNode: SHUTDOWN_MSG: > /************************************************************ > SHUTDOWN_MSG: Shutting down NameNode at name.server.ac.uk/161.74.12.97 > ************************************************************/ > > Thank you, > A Df > > ________________________________ > From: Harsh J <[email protected]> > To: A Df <[email protected]> > Sent: Wednesday, 10 August 2011, 15:13 > Subject: Re: Where is web interface in stand alone operation? > > A Df, > > On Wed, Aug 10, 2011 at 7:28 PM, A Df <[email protected]> wrote: >> >> Hello Harsh: >> See inline at * >> >> ________________________________ >> From: Harsh J <[email protected]> >> To: [email protected]; A Df <[email protected]> >> Sent: Wednesday, 10 August 2011, 14:44 >> Subject: Re: Where is web interface in stand alone operation? >> >> A Df, >> >> The web UIs are a feature of the daemons JobTracker and NameNode. In >> standalone/'local'/'file:///' modes, these daemons aren't run >> (actually, no daemon is run at all), and hence there would be no 'web' >> interface. >> >> *ok, but is there any other way to check the performance in this mode such >> as time to complete etc? I am trying to compare performance between the >> two. >> And also for the pseudo mode how would I change the ports for the web >> interface because I have to connect to a remote server which only allows >> certain ports to be accessed from the web? > > The ports Kai mentioned above are sourced from the configs: > dfs.http.address (hdfs-site.xml) and mapred.job.tracker.http.address > (mapred-site.xml). You can change them to bind to a host:port of your > preference. > > > -- > Harsh J > > > -- Harsh J
