log out from the user. log in again and see if it works. Warm Regards, Tariq https://mtariq.jux.com/ cloudfront.blogspot.com
On Mon, Mar 18, 2013 at 12:18 AM, Mohammad Tariq <[email protected]> wrote: > you can avoid the warning by setting the following prop to true in the > hadoop-env.sh file : > export HADOOP_HOME_WARN_SUPPRESS=true > > > > Warm Regards, > Tariq > https://mtariq.jux.com/ > cloudfront.blogspot.com > > > On Mon, Mar 18, 2013 at 12:07 AM, Mohammad Alkahtani < > [email protected]> wrote: > >> Thank you Mohammad >> I still get the same error with this msg >> >> localhost: Warning: $HADOOP_HOME is deprecated. >> I searched ~/.bashrc but only what I wrote is there. >> >> >> Mohammad Alkahtani >> P.O.Box 102275 >> Riyadh 11675 >> Saudi Arabia >> mobile: 00966 555 33 1717 >> >> >> On Sun, Mar 17, 2013 at 9:21 PM, Mohammad Tariq <[email protected]>wrote: >> >>> you can do that using these command : >>> >>> sudo gedit ~/.bashrc >>> >>> then go to the end of the file and add this line : >>> export HADOOP_HOME=/YOUR_FULL_HADOOP_PATH >>> >>> after that use it to freeze the changes : >>> source ~/.bashrc >>> >>> to check it : >>> echo $HADOOP_HOME >>> >>> This will permanently set your HADOOP_HOME. >>> >>> HTH >>> >>> >>> Warm Regards, >>> Tariq >>> https://mtariq.jux.com/ >>> cloudfront.blogspot.com >>> >>> >>> On Sun, Mar 17, 2013 at 11:46 PM, Mohammad Alkahtani < >>> [email protected]> wrote: >>> >>>> Hi Tariq, Could you please tell me how to set HADOOP_HOME because I >>>> don't find it in the hadoop-env.sh >>>> >>>> Thank you Shashwat >>>> this is the output and it is already configured but hadoop don't read >>>> the configuration from here. >>>> >>>> /usr/share/maven-repo/org/apache >>>> /commons/commons-parent/22/commons-parent-22-site.xml >>>> /usr/share/maven-repo/org/apache/commons/commons-parent/debian >>>> /commons-parent-debian-site.xml >>>> /usr/share/maven-repo/org/apache/apache/10/apache-10-site.xml >>>> /usr/share/maven-repo/org/apache/apache/debian/apache-debian-site.xml >>>> /usr/share/compiz/composite.xml >>>> /usr/share/hadoop/templates/conf/mapred-site.xml >>>> /usr/share/hadoop/templates/conf/core-site.xml >>>> /usr/share/hadoop/templates/conf/hdfs-site.xml >>>> >>>> Mohammad Alkahtani >>>> P.O.Box 102275 >>>> Riyadh 11675 >>>> Saudi Arabia >>>> mobile: 00966 555 33 1717 >>>> >>>> >>>> On Sun, Mar 17, 2013 at 9:07 PM, shashwat shriparv < >>>> [email protected]> wrote: >>>> >>>>> try >>>>> find / -type f -iname "*site.xml" >>>>> it will show you where ever those files are.. >>>>> >>>>> >>>>> >>>>> ∞ >>>>> Shashwat Shriparv >>>>> >>>>> >>>>> >>>>> On Sun, Mar 17, 2013 at 11:34 PM, Mohammad Alkahtani < >>>>> [email protected]> wrote: >>>>> >>>>>> The problem is I tried I read the configuration file by changing >>>>>> export HADOOP_CONF_DIR=${HADOOP_CONF_ >>>>>> DIR:-"/usr/shar/hadoop/templates/conf"} >>>>>> but I think Hadoop dosen't get the configration from this dir, I trid >>>>>> and searched the system for conf dir the only dir is this one which I >>>>>> changed. >>>>>> >>>>>> Mohammad Alkahtani >>>>>> P.O.Box 102275 >>>>>> Riyadh 11675 >>>>>> Saudi Arabia >>>>>> mobile: 00966 555 33 1717 >>>>>> >>>>>> >>>>>> On Sun, Mar 17, 2013 at 8:57 PM, shashwat shriparv < >>>>>> [email protected]> wrote: >>>>>> >>>>>>> Ye its is asking for file:/// instead of hdfs:// just check if it is >>>>>>> taking setting configuration from other location... >>>>>>> >>>>>>> >>>>>>> >>>>>>> ∞ >>>>>>> Shashwat Shriparv >>>>>>> >>>>>>> >>>>>>> >>>>>>> On Sun, Mar 17, 2013 at 11:07 PM, Luangsay Sourygna < >>>>>>> [email protected]> wrote: >>>>>>> >>>>>>>> Hi, >>>>>>>> >>>>>>>> What is the version of Hadoop you use? >>>>>>>> >>>>>>>> Try using fs.defaultFS instead of fs.default.name (see the list of >>>>>>>> all >>>>>>>> the deprecated properties here: >>>>>>>> >>>>>>>> http://hadoop.apache.org/docs/current/hadoop-project-dist/hadoop-common/DeprecatedProperties.html >>>>>>>> ). >>>>>>>> I remember I once had a similar error message and it was due to the >>>>>>>> change in properties names. >>>>>>>> >>>>>>>> Regards, >>>>>>>> >>>>>>>> Sourygna >>>>>>>> >>>>>>>> On Sun, Mar 17, 2013 at 2:32 PM, Mohammad Alkahtani >>>>>>>> <[email protected]> wrote: >>>>>>>> > Hi to all users of Hadoop, >>>>>>>> > >>>>>>>> > I installed Hadoop the .deb file on Ubuntu 12.04 but I might >>>>>>>> could not >>>>>>>> > configure it right. The conf dir is under templates in >>>>>>>> /usr/shar/hadoop. I >>>>>>>> > edit the core-site.xml, mapred-site.xml files to give >>>>>>>> > <property> >>>>>>>> > <name>fs.default.name</name> >>>>>>>> > <value>hdfs://localhost:9000</value> >>>>>>>> > </property> >>>>>>>> > and for mapred >>>>>>>> > <property> >>>>>>>> > <name>mapred.job.tracker</name> >>>>>>>> > <value>localhost:9001</value> >>>>>>>> > </property> >>>>>>>> > >>>>>>>> > but i get these errors, I assume that there is problem, Hadoop >>>>>>>> cannot read >>>>>>>> > the configuration file. >>>>>>>> > I chaned the hadoop-env.sh to >>>>>>>> > export >>>>>>>> HADOOP_CONF_DIR=${HADOOP_CONF_DIR:-"/usr/shar/hadoop/templates/conf"} >>>>>>>> > but dosen't solve the problem. >>>>>>>> > >>>>>>>> > ERROR org.apache.hadoop.hdfs.server.datanode.DataNode: >>>>>>>> > java.lang.IllegalArgumentException: Does not contain a valid >>>>>>>> host:port >>>>>>>> > authority: file:/// at >>>>>>>> > >>>>>>>> org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:164) at >>>>>>>> > >>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:201) >>>>>>>> > at >>>>>>>> > >>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:231) >>>>>>>> > at >>>>>>>> > >>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getServiceAddress(NameNode.java:225) >>>>>>>> > at >>>>>>>> > >>>>>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:347) >>>>>>>> > at >>>>>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.(DataNode.java:309) at >>>>>>>> > >>>>>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1651) >>>>>>>> > at >>>>>>>> > >>>>>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1590) >>>>>>>> > at >>>>>>>> > >>>>>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:1608) >>>>>>>> > at >>>>>>>> > >>>>>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.secureMain(DataNode.java:1734) >>>>>>>> > at >>>>>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:1751) >>>>>>>> > >>>>>>>> > ________________________________ >>>>>>>> > >>>>>>>> > FATAL org.apache.hadoop.mapred.JobTracker: >>>>>>>> > java.lang.IllegalArgumentException: Does not contain a valid >>>>>>>> host:port >>>>>>>> > authority: local at >>>>>>>> > >>>>>>>> org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:164) at >>>>>>>> > >>>>>>>> org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:130) at >>>>>>>> > >>>>>>>> org.apache.hadoop.mapred.JobTracker.getAddress(JobTracker.java:2312) at >>>>>>>> > org.apache.hadoop.mapred.JobTracker.(JobTracker.java:2070) at >>>>>>>> > org.apache.hadoop.mapred.JobTracker.(JobTracker.java:1889) at >>>>>>>> > org.apache.hadoop.mapred.JobTracker.(JobTracker.java:1883) at >>>>>>>> > >>>>>>>> org.apache.hadoop.mapred.JobTracker.startTracker(JobTracker.java:312) >>>>>>>> at >>>>>>>> > >>>>>>>> org.apache.hadoop.mapred.JobTracker.startTracker(JobTracker.java:303) >>>>>>>> at >>>>>>>> > >>>>>>>> org.apache.hadoop.mapred.JobTracker.startTracker(JobTracker.java:298) >>>>>>>> > at org.apache.hadoop.mapred.JobTracker.main(JobTracker.java:4791) >>>>>>>> > >>>>>>>> > ________________________________ >>>>>>>> > >>>>>>>> > ERROR org.apache.hadoop.hdfs.server.namenode.NameNode: >>>>>>>> > java.lang.IllegalArgumentException: Does not contain a valid >>>>>>>> host:port >>>>>>>> > authority: file:/// at >>>>>>>> > >>>>>>>> org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:164) at >>>>>>>> > >>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:201) >>>>>>>> > at >>>>>>>> > >>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:231) >>>>>>>> > at >>>>>>>> > >>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:265) >>>>>>>> > at >>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.(NameNode.java:536) at >>>>>>>> > >>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1410) >>>>>>>> > at >>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1419) >>>>>>>> > >>>>>>>> > ________________________________ >>>>>>>> > >>>>>>>> > Exception in thread "main" java.lang.IllegalArgumentException: >>>>>>>> Does not >>>>>>>> > contain a valid host:port authority: file:/// at >>>>>>>> > >>>>>>>> org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:164) at >>>>>>>> > >>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:201) >>>>>>>> > at >>>>>>>> > >>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:231) >>>>>>>> > at >>>>>>>> > >>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getServiceAddress(NameNode.java:225) >>>>>>>> > at >>>>>>>> > >>>>>>>> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.initialize(SecondaryNameNode.java:167) >>>>>>>> > at >>>>>>>> > >>>>>>>> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.(SecondaryNameNode.java:135) >>>>>>>> > at >>>>>>>> > >>>>>>>> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.main(SecondaryNameNode.java:650) >>>>>>>> > >>>>>>>> > ________________________________ >>>>>>>> > >>>>>>>> > ERROR org.apache.hadoop.mapred.TaskTracker: Can not start task >>>>>>>> tracker >>>>>>>> > because java.lang.IllegalArgumentException: Does not contain a >>>>>>>> valid >>>>>>>> > host:port authority: local at >>>>>>>> > >>>>>>>> org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:164) at >>>>>>>> > >>>>>>>> org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:130) at >>>>>>>> > >>>>>>>> org.apache.hadoop.mapred.JobTracker.getAddress(JobTracker.java:2312) at >>>>>>>> > org.apache.hadoop.mapred.TaskTracker.(TaskTracker.java:1532) at >>>>>>>> > org.apache.hadoop.mapred.TaskTracker.main(TaskTracker.java:3906) >>>>>>>> > >>>>>>>> > >>>>>>>> > Regards, >>>>>>>> > Mohammad Alkahtani >>>>>>>> >>>>>>> >>>>>>> >>>>>> >>>>> >>>> >>> >> >
