you can do that using these command : sudo gedit ~/.bashrc
then go to the end of the file and add this line : export HADOOP_HOME=/YOUR_FULL_HADOOP_PATH after that use it to freeze the changes : source ~/.bashrc to check it : echo $HADOOP_HOME This will permanently set your HADOOP_HOME. HTH Warm Regards, Tariq https://mtariq.jux.com/ cloudfront.blogspot.com On Sun, Mar 17, 2013 at 11:46 PM, Mohammad Alkahtani <[email protected]>wrote: > Hi Tariq, Could you please tell me how to set HADOOP_HOME because I don't > find it in the hadoop-env.sh > > Thank you Shashwat > this is the output and it is already configured but hadoop don't read the > configuration from here. > > /usr/share/maven-repo/org/apache > /commons/commons-parent/22/commons-parent-22-site.xml > /usr/share/maven-repo/org/apache/commons/commons-parent/debian > /commons-parent-debian-site.xml > /usr/share/maven-repo/org/apache/apache/10/apache-10-site.xml > /usr/share/maven-repo/org/apache/apache/debian/apache-debian-site.xml > /usr/share/compiz/composite.xml > /usr/share/hadoop/templates/conf/mapred-site.xml > /usr/share/hadoop/templates/conf/core-site.xml > /usr/share/hadoop/templates/conf/hdfs-site.xml > > Mohammad Alkahtani > P.O.Box 102275 > Riyadh 11675 > Saudi Arabia > mobile: 00966 555 33 1717 > > > On Sun, Mar 17, 2013 at 9:07 PM, shashwat shriparv < > [email protected]> wrote: > >> try >> find / -type f -iname "*site.xml" >> it will show you where ever those files are.. >> >> >> >> ∞ >> Shashwat Shriparv >> >> >> >> On Sun, Mar 17, 2013 at 11:34 PM, Mohammad Alkahtani < >> [email protected]> wrote: >> >>> The problem is I tried I read the configuration file by changing >>> export HADOOP_CONF_DIR=${HADOOP_CONF_ >>> DIR:-"/usr/shar/hadoop/templates/conf"} >>> but I think Hadoop dosen't get the configration from this dir, I trid >>> and searched the system for conf dir the only dir is this one which I >>> changed. >>> >>> Mohammad Alkahtani >>> P.O.Box 102275 >>> Riyadh 11675 >>> Saudi Arabia >>> mobile: 00966 555 33 1717 >>> >>> >>> On Sun, Mar 17, 2013 at 8:57 PM, shashwat shriparv < >>> [email protected]> wrote: >>> >>>> Ye its is asking for file:/// instead of hdfs:// just check if it is >>>> taking setting configuration from other location... >>>> >>>> >>>> >>>> ∞ >>>> Shashwat Shriparv >>>> >>>> >>>> >>>> On Sun, Mar 17, 2013 at 11:07 PM, Luangsay Sourygna <[email protected] >>>> > wrote: >>>> >>>>> Hi, >>>>> >>>>> What is the version of Hadoop you use? >>>>> >>>>> Try using fs.defaultFS instead of fs.default.name (see the list of all >>>>> the deprecated properties here: >>>>> >>>>> http://hadoop.apache.org/docs/current/hadoop-project-dist/hadoop-common/DeprecatedProperties.html >>>>> ). >>>>> I remember I once had a similar error message and it was due to the >>>>> change in properties names. >>>>> >>>>> Regards, >>>>> >>>>> Sourygna >>>>> >>>>> On Sun, Mar 17, 2013 at 2:32 PM, Mohammad Alkahtani >>>>> <[email protected]> wrote: >>>>> > Hi to all users of Hadoop, >>>>> > >>>>> > I installed Hadoop the .deb file on Ubuntu 12.04 but I might could >>>>> not >>>>> > configure it right. The conf dir is under templates in >>>>> /usr/shar/hadoop. I >>>>> > edit the core-site.xml, mapred-site.xml files to give >>>>> > <property> >>>>> > <name>fs.default.name</name> >>>>> > <value>hdfs://localhost:9000</value> >>>>> > </property> >>>>> > and for mapred >>>>> > <property> >>>>> > <name>mapred.job.tracker</name> >>>>> > <value>localhost:9001</value> >>>>> > </property> >>>>> > >>>>> > but i get these errors, I assume that there is problem, Hadoop >>>>> cannot read >>>>> > the configuration file. >>>>> > I chaned the hadoop-env.sh to >>>>> > export >>>>> HADOOP_CONF_DIR=${HADOOP_CONF_DIR:-"/usr/shar/hadoop/templates/conf"} >>>>> > but dosen't solve the problem. >>>>> > >>>>> > ERROR org.apache.hadoop.hdfs.server.datanode.DataNode: >>>>> > java.lang.IllegalArgumentException: Does not contain a valid >>>>> host:port >>>>> > authority: file:/// at >>>>> > org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:164) at >>>>> > >>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:201) >>>>> > at >>>>> > >>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:231) >>>>> > at >>>>> > >>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getServiceAddress(NameNode.java:225) >>>>> > at >>>>> > >>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:347) >>>>> > at >>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.(DataNode.java:309) at >>>>> > >>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1651) >>>>> > at >>>>> > >>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1590) >>>>> > at >>>>> > >>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:1608) >>>>> > at >>>>> > >>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.secureMain(DataNode.java:1734) >>>>> > at >>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:1751) >>>>> > >>>>> > ________________________________ >>>>> > >>>>> > FATAL org.apache.hadoop.mapred.JobTracker: >>>>> > java.lang.IllegalArgumentException: Does not contain a valid >>>>> host:port >>>>> > authority: local at >>>>> > org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:164) at >>>>> > org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:130) at >>>>> > org.apache.hadoop.mapred.JobTracker.getAddress(JobTracker.java:2312) >>>>> at >>>>> > org.apache.hadoop.mapred.JobTracker.(JobTracker.java:2070) at >>>>> > org.apache.hadoop.mapred.JobTracker.(JobTracker.java:1889) at >>>>> > org.apache.hadoop.mapred.JobTracker.(JobTracker.java:1883) at >>>>> > >>>>> org.apache.hadoop.mapred.JobTracker.startTracker(JobTracker.java:312) at >>>>> > >>>>> org.apache.hadoop.mapred.JobTracker.startTracker(JobTracker.java:303) at >>>>> > org.apache.hadoop.mapred.JobTracker.startTracker(JobTracker.java:298) >>>>> > at org.apache.hadoop.mapred.JobTracker.main(JobTracker.java:4791) >>>>> > >>>>> > ________________________________ >>>>> > >>>>> > ERROR org.apache.hadoop.hdfs.server.namenode.NameNode: >>>>> > java.lang.IllegalArgumentException: Does not contain a valid >>>>> host:port >>>>> > authority: file:/// at >>>>> > org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:164) at >>>>> > >>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:201) >>>>> > at >>>>> > >>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:231) >>>>> > at >>>>> > >>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:265) >>>>> > at >>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.(NameNode.java:536) at >>>>> > >>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1410) >>>>> > at >>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1419) >>>>> > >>>>> > ________________________________ >>>>> > >>>>> > Exception in thread "main" java.lang.IllegalArgumentException: Does >>>>> not >>>>> > contain a valid host:port authority: file:/// at >>>>> > org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:164) at >>>>> > >>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:201) >>>>> > at >>>>> > >>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:231) >>>>> > at >>>>> > >>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getServiceAddress(NameNode.java:225) >>>>> > at >>>>> > >>>>> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.initialize(SecondaryNameNode.java:167) >>>>> > at >>>>> > >>>>> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.(SecondaryNameNode.java:135) >>>>> > at >>>>> > >>>>> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.main(SecondaryNameNode.java:650) >>>>> > >>>>> > ________________________________ >>>>> > >>>>> > ERROR org.apache.hadoop.mapred.TaskTracker: Can not start task >>>>> tracker >>>>> > because java.lang.IllegalArgumentException: Does not contain a valid >>>>> > host:port authority: local at >>>>> > org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:164) at >>>>> > org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:130) at >>>>> > org.apache.hadoop.mapred.JobTracker.getAddress(JobTracker.java:2312) >>>>> at >>>>> > org.apache.hadoop.mapred.TaskTracker.(TaskTracker.java:1532) at >>>>> > org.apache.hadoop.mapred.TaskTracker.main(TaskTracker.java:3906) >>>>> > >>>>> > >>>>> > Regards, >>>>> > Mohammad Alkahtani >>>>> >>>> >>>> >>> >> >
