Brahma, I found the problem. I had installed hadoop from Ubuntu's distribution. Then I installed hadoop from the Apache tar file and configured it according to Michael Noll's tutorial. My mistake was that I should have uninstalled the Ubuntu distribution first.
Thank you for your assistance. Jeff Silverman On Wed, Jun 27, 2012 at 10:05 PM, Brahma Reddy Battula < brahmareddy.batt...@huawei.com> wrote: > HI Jeff, > > *Do you have another suggestion? I think the problem is that somewhere > I have URL that is in the form file:// instead of hdfs://. I have done > some fgrep'ing and I see several possibilities, but nothing jumps out at me. > * > > Mostly it related classpath issue since it is not considering what ever > you configured in core-site.xml..Please check class path..May be hadoop > conf you might have put in classpath which different from where ever you > are starting.. > > If NameNode and Datanode are in same machine,they also wn't start..I > think they are in different machine.. > > Please correct me If I am wrong.. > > > Thanks And Regards > > Brahma Reddy > > > ------------------------------ > *From:* Jeffrey Silverman [jeffsilver...@google.com] > *Sent:* Thursday, June 28, 2012 2:14 AM > *To:* hdfs-user@hadoop.apache.org > *Subject:* Re: Problems starting secondarynamenode in hadoop 1.0.3 > > > Varun, > > I tried what you suggested and I am still having the same problem: > > hduser@master:/usr/local/hadoop-1.0.3/conf$ tail > /var/log/hadoop/hduser/hadoop-hduser-secondarynamenode-master.out > Exception in thread "main" java.lang.IllegalArgumentException: Does not > contain a valid host:port authority: file:/// > at org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:162) > at > org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:198) > at > org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:228) > at > org.apache.hadoop.hdfs.server.namenode.NameNode.getServiceAddress(NameNode.java:222) > at > org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.initialize(SecondaryNameNode.java:161) > at > org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.<init>(SecondaryNameNode.java:129) > at > org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.main(SecondaryNameNode.java:567) > hduser@master:/usr/local/hadoop-1.0.3/conf$ > > > hduser@master:/usr/local/hadoop-1.0.3/conf$ cat core-site.xml > <?xml version="1.0"?> > <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> > > <!-- Put site-specific property overrides in this file. --> > > <configuration> > <property> > <name>hadoop.tmp.dir</name> > <value>/home/hduser/tmp</value> > <description>A base for other temporary directories.</description> > </property> > > <property> > <name>fs.default.name</name> > <!-- The instructions in > http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-single-node-cluster/#conf-site-xml > says that the value should be hdfs://localhost:54310 however I got an > answer on the hadoop mailing list that said to explicitly > name the host --> > <value>hdfs://master:54310</value> > <description>The name of the default file system. A URI whose > scheme and authority determine the FileSystem implementation. The > uri's scheme determines the config property (fs.SCHEME.impl) naming > the FileSystem implementation class. The uri's authority is used to > determine the host, port, etc. for a filesystem.</description> > </property> > </configuration> > hduser@master:/usr/local/hadoop-1.0.3/conf$ > > Do you have another suggestion? I think the problem is that somewhere I > have URL that is in the form file:// instead of hdfs://. I have done some > fgrep'ing and I see several possibilities, but nothing jumps out at me. > > Thank you > > > Jeff > > > On Tue, Jun 26, 2012 at 7:44 PM, varun kumar <varun....@gmail.com> wrote: > >> Hi Jeff, >> >> Instead of localhost,mention the host-name of Primary namenod. >> >> >> On Wed, Jun 27, 2012 at 3:46 AM, Jeffrey Silverman < >> jeffsilver...@google.com> wrote: >> >>> I am working with hadoop for the first time, and I am following >>> instructions at >>> http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-single-node-cluster/ >>> >>> >>> I am having problems starting the secondarynamenode daemon. The error >>> message in >>> /var/log/hadoop/hduser/hadoop-hduser-secondarynamenode-master.out is >>> >>> Exception in thread "main" java.lang.IllegalArgumentException: Does >>> not contain a valid host:port authority: file:/// >>> at >>> org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:162) >>> at >>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:198) >>> at >>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:228) >>> at >>> org.apache.hadoop.hdfs.server.namenode.NameNode.getServiceAddress(NameNode.java:222) >>> at >>> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.initialize(SecondaryNameNode.java:161) >>> at >>> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.<init>(SecondaryNameNode.java:129) >>> at >>> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.main(SecondaryNameNode.java:567) >>> >>> >>> >>> I googled the error message and came across HDFS-2515, which says that >>> I might get that error message if the fs.default.name property name had >>> an incorrect value, but I think my value is okay. >>> >>> My core-site.xml file is : >>> >>> <?xml version="1.0"?> >>> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> >>> >>> <!-- Put site-specific property overrides in this file. --> >>> >>> <configuration> >>> <property> >>> <name>hadoop.tmp.dir</name> >>> <value>/home/hduser/tmp</value> >>> <description>A base for other temporary directories.</description> >>> </property> >>> >>> <property> >>> <name>fs.default.name</name> >>> <value>hdfs://localhost:54310</value> >>> <description>The name of the default file system. A URI whose >>> scheme and authority determine the FileSystem implementation. The >>> uri's scheme determines the config property (fs.SCHEME.impl) naming >>> the FileSystem implementation class. The uri's authority is used to >>> determine the host, port, etc. for a filesystem.</description> >>> </property> >>> </configuration> >>> ~ >>> >>> Does anybody have a suggestion for how to further troubleshoot this >>> problem, please? >>> >>> >>> Thank you, >>> >>> >>> Jeff Silverman >>> >>> >> >> >> >> -- >> Regards, >> Varun Kumar.P >> >> >