Hi Tariq, Could you please tell me how to set HADOOP_HOME because I don't find it in the hadoop-env.sh
Thank you Shashwat this is the output and it is already configured but hadoop don't read the configuration from here. /usr/share/maven-repo/org/apache /commons/commons-parent/22/commons-parent-22-site.xml /usr/share/maven-repo/org/apache/commons/commons-parent/debian /commons-parent-debian-site.xml /usr/share/maven-repo/org/apache/apache/10/apache-10-site.xml /usr/share/maven-repo/org/apache/apache/debian/apache-debian-site.xml /usr/share/compiz/composite.xml /usr/share/hadoop/templates/conf/mapred-site.xml /usr/share/hadoop/templates/conf/core-site.xml /usr/share/hadoop/templates/conf/hdfs-site.xml Mohammad Alkahtani P.O.Box 102275 Riyadh 11675 Saudi Arabia mobile: 00966 555 33 1717 On Sun, Mar 17, 2013 at 9:07 PM, shashwat shriparv < [email protected]> wrote: > try > find / -type f -iname "*site.xml" > it will show you where ever those files are.. > > > > ∞ > Shashwat Shriparv > > > > On Sun, Mar 17, 2013 at 11:34 PM, Mohammad Alkahtani < > [email protected]> wrote: > >> The problem is I tried I read the configuration file by changing >> export HADOOP_CONF_DIR=${HADOOP_CONF_ >> DIR:-"/usr/shar/hadoop/templates/conf"} >> but I think Hadoop dosen't get the configration from this dir, I trid and >> searched the system for conf dir the only dir is this one which I changed. >> >> Mohammad Alkahtani >> P.O.Box 102275 >> Riyadh 11675 >> Saudi Arabia >> mobile: 00966 555 33 1717 >> >> >> On Sun, Mar 17, 2013 at 8:57 PM, shashwat shriparv < >> [email protected]> wrote: >> >>> Ye its is asking for file:/// instead of hdfs:// just check if it is >>> taking setting configuration from other location... >>> >>> >>> >>> ∞ >>> Shashwat Shriparv >>> >>> >>> >>> On Sun, Mar 17, 2013 at 11:07 PM, Luangsay Sourygna >>> <[email protected]>wrote: >>> >>>> Hi, >>>> >>>> What is the version of Hadoop you use? >>>> >>>> Try using fs.defaultFS instead of fs.default.name (see the list of all >>>> the deprecated properties here: >>>> >>>> http://hadoop.apache.org/docs/current/hadoop-project-dist/hadoop-common/DeprecatedProperties.html >>>> ). >>>> I remember I once had a similar error message and it was due to the >>>> change in properties names. >>>> >>>> Regards, >>>> >>>> Sourygna >>>> >>>> On Sun, Mar 17, 2013 at 2:32 PM, Mohammad Alkahtani >>>> <[email protected]> wrote: >>>> > Hi to all users of Hadoop, >>>> > >>>> > I installed Hadoop the .deb file on Ubuntu 12.04 but I might could not >>>> > configure it right. The conf dir is under templates in >>>> /usr/shar/hadoop. I >>>> > edit the core-site.xml, mapred-site.xml files to give >>>> > <property> >>>> > <name>fs.default.name</name> >>>> > <value>hdfs://localhost:9000</value> >>>> > </property> >>>> > and for mapred >>>> > <property> >>>> > <name>mapred.job.tracker</name> >>>> > <value>localhost:9001</value> >>>> > </property> >>>> > >>>> > but i get these errors, I assume that there is problem, Hadoop cannot >>>> read >>>> > the configuration file. >>>> > I chaned the hadoop-env.sh to >>>> > export >>>> HADOOP_CONF_DIR=${HADOOP_CONF_DIR:-"/usr/shar/hadoop/templates/conf"} >>>> > but dosen't solve the problem. >>>> > >>>> > ERROR org.apache.hadoop.hdfs.server.datanode.DataNode: >>>> > java.lang.IllegalArgumentException: Does not contain a valid host:port >>>> > authority: file:/// at >>>> > org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:164) at >>>> > >>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:201) >>>> > at >>>> > >>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:231) >>>> > at >>>> > >>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getServiceAddress(NameNode.java:225) >>>> > at >>>> > >>>> org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:347) >>>> > at >>>> org.apache.hadoop.hdfs.server.datanode.DataNode.(DataNode.java:309) at >>>> > >>>> org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1651) >>>> > at >>>> > >>>> org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1590) >>>> > at >>>> > >>>> org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:1608) >>>> > at >>>> > >>>> org.apache.hadoop.hdfs.server.datanode.DataNode.secureMain(DataNode.java:1734) >>>> > at >>>> org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:1751) >>>> > >>>> > ________________________________ >>>> > >>>> > FATAL org.apache.hadoop.mapred.JobTracker: >>>> > java.lang.IllegalArgumentException: Does not contain a valid host:port >>>> > authority: local at >>>> > org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:164) at >>>> > org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:130) at >>>> > org.apache.hadoop.mapred.JobTracker.getAddress(JobTracker.java:2312) >>>> at >>>> > org.apache.hadoop.mapred.JobTracker.(JobTracker.java:2070) at >>>> > org.apache.hadoop.mapred.JobTracker.(JobTracker.java:1889) at >>>> > org.apache.hadoop.mapred.JobTracker.(JobTracker.java:1883) at >>>> > org.apache.hadoop.mapred.JobTracker.startTracker(JobTracker.java:312) >>>> at >>>> > org.apache.hadoop.mapred.JobTracker.startTracker(JobTracker.java:303) >>>> at >>>> > org.apache.hadoop.mapred.JobTracker.startTracker(JobTracker.java:298) >>>> > at org.apache.hadoop.mapred.JobTracker.main(JobTracker.java:4791) >>>> > >>>> > ________________________________ >>>> > >>>> > ERROR org.apache.hadoop.hdfs.server.namenode.NameNode: >>>> > java.lang.IllegalArgumentException: Does not contain a valid host:port >>>> > authority: file:/// at >>>> > org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:164) at >>>> > >>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:201) >>>> > at >>>> > >>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:231) >>>> > at >>>> > >>>> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:265) >>>> > at >>>> org.apache.hadoop.hdfs.server.namenode.NameNode.(NameNode.java:536) at >>>> > >>>> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1410) >>>> > at >>>> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1419) >>>> > >>>> > ________________________________ >>>> > >>>> > Exception in thread "main" java.lang.IllegalArgumentException: Does >>>> not >>>> > contain a valid host:port authority: file:/// at >>>> > org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:164) at >>>> > >>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:201) >>>> > at >>>> > >>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:231) >>>> > at >>>> > >>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getServiceAddress(NameNode.java:225) >>>> > at >>>> > >>>> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.initialize(SecondaryNameNode.java:167) >>>> > at >>>> > >>>> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.(SecondaryNameNode.java:135) >>>> > at >>>> > >>>> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.main(SecondaryNameNode.java:650) >>>> > >>>> > ________________________________ >>>> > >>>> > ERROR org.apache.hadoop.mapred.TaskTracker: Can not start task tracker >>>> > because java.lang.IllegalArgumentException: Does not contain a valid >>>> > host:port authority: local at >>>> > org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:164) at >>>> > org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:130) at >>>> > org.apache.hadoop.mapred.JobTracker.getAddress(JobTracker.java:2312) >>>> at >>>> > org.apache.hadoop.mapred.TaskTracker.(TaskTracker.java:1532) at >>>> > org.apache.hadoop.mapred.TaskTracker.main(TaskTracker.java:3906) >>>> > >>>> > >>>> > Regards, >>>> > Mohammad Alkahtani >>>> >>> >>> >> >
