you have to use upper case 'HADOOP_HOME' (don't mind if it's a typo). do
you proper permission to read these files?

Warm Regards,
Tariq
https://mtariq.jux.com/
cloudfront.blogspot.com


On Mon, Mar 18, 2013 at 1:00 AM, Mohammad Alkahtani
<[email protected]>wrote:

> The files that in hadoop-x.x.x/bin in the /usr/bin di, I tried to set the
> Hadoop_Home to /usr but still get the errors, I tried /etc/hadoop also I
> got the error.
>
> Mohammad Alkahtani
> P.O.Box 102275
> Riyadh 11675
> Saudi Arabia
> mobile: 00966 555 33 1717
>
>
> On Sun, Mar 17, 2013 at 10:15 PM, Mohammad Tariq <[email protected]>wrote:
>
>> set these properties in the configuration files present in your /etc
>> directory. and HADOOP_HOME is the parent directory of the hadoop bin
>> directory that holds the Hadoop scripts. so, set that accordingly in
>> .bashrc file.
>>
>> Warm Regards,
>> Tariq
>> https://mtariq.jux.com/
>> cloudfront.blogspot.com
>>
>>
>> On Mon, Mar 18, 2013 at 12:35 AM, Mohammad Alkahtani <
>> [email protected]> wrote:
>>
>>> Thank you Mohammad Tariq
>>>
>>> Mohammad Alkahtani
>>> P.O.Box 102275
>>> Riyadh 11675
>>> Saudi Arabia
>>> mobile: 00966 555 33 1717
>>>
>>>
>>> On Sun, Mar 17, 2013 at 10:04 PM, Mohammad Alkahtani <
>>> [email protected]> wrote:
>>>
>>>> I tried all of the hadoop home dirs but didn't worke
>>>>
>>>> Mohammad Alkahtani
>>>> P.O.Box 102275
>>>> Riyadh 11675
>>>> Saudi Arabia
>>>> mobile: 00966 555 33 1717
>>>>
>>>>
>>>> On Sun, Mar 17, 2013 at 9:57 PM, Mohammad Alkahtani <
>>>> [email protected]> wrote:
>>>>
>>>>> OK what the Hadoop home should be in ubuntu because the binary files
>>>>> in /usr/bin
>>>>> the hadoop-env.sh and othe xml file in /etc/hadoop
>>>>> the conf files in /usr/share/hadoop/templates/conf
>>>>>
>>>>> shall I use /usr as hadoop path because it is the dir that contain the
>>>>> bin files
>>>>>
>>>>> Mohammad Alkahtani
>>>>> P.O.Box 102275
>>>>> Riyadh 11675
>>>>> Saudi Arabia
>>>>> mobile: 00966 555 33 1717
>>>>>
>>>>>
>>>>> On Sun, Mar 17, 2013 at 9:50 PM, Mohammad Tariq <[email protected]>wrote:
>>>>>
>>>>>> log out from the user. log in again and see if it works.
>>>>>>
>>>>>> Warm Regards,
>>>>>> Tariq
>>>>>> https://mtariq.jux.com/
>>>>>> cloudfront.blogspot.com
>>>>>>
>>>>>>
>>>>>> On Mon, Mar 18, 2013 at 12:18 AM, Mohammad Tariq 
>>>>>> <[email protected]>wrote:
>>>>>>
>>>>>>> you can avoid the warning by setting the following prop to true in
>>>>>>> the hadoop-env.sh file :
>>>>>>> export HADOOP_HOME_WARN_SUPPRESS=true
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> Warm Regards,
>>>>>>> Tariq
>>>>>>> https://mtariq.jux.com/
>>>>>>> cloudfront.blogspot.com
>>>>>>>
>>>>>>>
>>>>>>> On Mon, Mar 18, 2013 at 12:07 AM, Mohammad Alkahtani <
>>>>>>> [email protected]> wrote:
>>>>>>>
>>>>>>>> Thank you Mohammad
>>>>>>>> I still get the same error with this msg
>>>>>>>>
>>>>>>>> localhost: Warning: $HADOOP_HOME is deprecated.
>>>>>>>> I searched ~/.bashrc but only what I wrote is there.
>>>>>>>>
>>>>>>>>
>>>>>>>> Mohammad Alkahtani
>>>>>>>> P.O.Box 102275
>>>>>>>> Riyadh 11675
>>>>>>>> Saudi Arabia
>>>>>>>> mobile: 00966 555 33 1717
>>>>>>>>
>>>>>>>>
>>>>>>>> On Sun, Mar 17, 2013 at 9:21 PM, Mohammad Tariq <[email protected]
>>>>>>>> > wrote:
>>>>>>>>
>>>>>>>>> you can do that using these command :
>>>>>>>>>
>>>>>>>>> sudo gedit ~/.bashrc
>>>>>>>>>
>>>>>>>>> then go to the end of the file and add this line :
>>>>>>>>> export HADOOP_HOME=/YOUR_FULL_HADOOP_PATH
>>>>>>>>>
>>>>>>>>> after that use it to freeze the changes :
>>>>>>>>> source ~/.bashrc
>>>>>>>>>
>>>>>>>>> to check it :
>>>>>>>>> echo $HADOOP_HOME
>>>>>>>>>
>>>>>>>>> This will permanently set your HADOOP_HOME.
>>>>>>>>>
>>>>>>>>> HTH
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> Warm Regards,
>>>>>>>>> Tariq
>>>>>>>>> https://mtariq.jux.com/
>>>>>>>>> cloudfront.blogspot.com
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On Sun, Mar 17, 2013 at 11:46 PM, Mohammad Alkahtani <
>>>>>>>>> [email protected]> wrote:
>>>>>>>>>
>>>>>>>>>> Hi Tariq, Could you please tell me how to set HADOOP_HOME
>>>>>>>>>> because I don't find it in the hadoop-env.sh
>>>>>>>>>>
>>>>>>>>>> Thank you Shashwat
>>>>>>>>>> this is the output and it is already configured but hadoop don't
>>>>>>>>>> read the configuration from here.
>>>>>>>>>>
>>>>>>>>>> /usr/share/maven-repo/org/apache
>>>>>>>>>> /commons/commons-parent/22/commons-parent-22-site.xml
>>>>>>>>>> /usr/share/maven-repo/org/apache/commons/commons-parent/debian
>>>>>>>>>> /commons-parent-debian-site.xml
>>>>>>>>>> /usr/share/maven-repo/org/apache/apache/10/apache-10-site.xml
>>>>>>>>>> /usr/share/maven-repo/org/apache/apache/debian/apache-debian
>>>>>>>>>> -site.xml
>>>>>>>>>> /usr/share/compiz/composite.xml
>>>>>>>>>> /usr/share/hadoop/templates/conf/mapred-site.xml
>>>>>>>>>> /usr/share/hadoop/templates/conf/core-site.xml
>>>>>>>>>> /usr/share/hadoop/templates/conf/hdfs-site.xml
>>>>>>>>>>
>>>>>>>>>> Mohammad Alkahtani
>>>>>>>>>> P.O.Box 102275
>>>>>>>>>> Riyadh 11675
>>>>>>>>>> Saudi Arabia
>>>>>>>>>> mobile: 00966 555 33 1717
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> On Sun, Mar 17, 2013 at 9:07 PM, shashwat shriparv <
>>>>>>>>>> [email protected]> wrote:
>>>>>>>>>>
>>>>>>>>>>> try
>>>>>>>>>>> find / -type f -iname "*site.xml"
>>>>>>>>>>> it will show you where ever those files are..
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> ∞
>>>>>>>>>>> Shashwat Shriparv
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> On Sun, Mar 17, 2013 at 11:34 PM, Mohammad Alkahtani <
>>>>>>>>>>> [email protected]> wrote:
>>>>>>>>>>>
>>>>>>>>>>>> The problem is I tried I read the configuration file by
>>>>>>>>>>>> changing
>>>>>>>>>>>> export HADOOP_CONF_DIR=${HADOOP_CONF_
>>>>>>>>>>>> DIR:-"/usr/shar/hadoop/templates/conf"}
>>>>>>>>>>>> but I think Hadoop dosen't get the configration from this dir,
>>>>>>>>>>>> I trid and searched the system for conf dir the only dir is this 
>>>>>>>>>>>> one which
>>>>>>>>>>>> I changed.
>>>>>>>>>>>>
>>>>>>>>>>>> Mohammad Alkahtani
>>>>>>>>>>>> P.O.Box 102275
>>>>>>>>>>>> Riyadh 11675
>>>>>>>>>>>> Saudi Arabia
>>>>>>>>>>>> mobile: 00966 555 33 1717
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> On Sun, Mar 17, 2013 at 8:57 PM, shashwat shriparv <
>>>>>>>>>>>> [email protected]> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>>> Ye its is asking for file:/// instead of hdfs:// just check if
>>>>>>>>>>>>> it is taking setting configuration from other location...
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> ∞
>>>>>>>>>>>>> Shashwat Shriparv
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> On Sun, Mar 17, 2013 at 11:07 PM, Luangsay Sourygna <
>>>>>>>>>>>>> [email protected]> wrote:
>>>>>>>>>>>>>
>>>>>>>>>>>>>> Hi,
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> What is the version of Hadoop you use?
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Try using fs.defaultFS instead of fs.default.name (see the
>>>>>>>>>>>>>> list of all
>>>>>>>>>>>>>> the deprecated properties here:
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> http://hadoop.apache.org/docs/current/hadoop-project-dist/hadoop-common/DeprecatedProperties.html
>>>>>>>>>>>>>> ).
>>>>>>>>>>>>>> I remember I once had a similar error message and it was due
>>>>>>>>>>>>>> to the
>>>>>>>>>>>>>> change in properties names.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Regards,
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Sourygna
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> On Sun, Mar 17, 2013 at 2:32 PM, Mohammad Alkahtani
>>>>>>>>>>>>>> <[email protected]> wrote:
>>>>>>>>>>>>>> > Hi to all users of Hadoop,
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> > I installed Hadoop the .deb file on Ubuntu 12.04 but I
>>>>>>>>>>>>>> might could not
>>>>>>>>>>>>>> > configure it right. The conf dir is under templates in
>>>>>>>>>>>>>> /usr/shar/hadoop. I
>>>>>>>>>>>>>> > edit the core-site.xml, mapred-site.xml files to give
>>>>>>>>>>>>>> > <property>
>>>>>>>>>>>>>> > <name>fs.default.name</name>
>>>>>>>>>>>>>> > <value>hdfs://localhost:9000</value>
>>>>>>>>>>>>>> > </property>
>>>>>>>>>>>>>> > and for mapred
>>>>>>>>>>>>>> > <property>
>>>>>>>>>>>>>> > <name>mapred.job.tracker</name>
>>>>>>>>>>>>>> > <value>localhost:9001</value>
>>>>>>>>>>>>>> > </property>
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> > but i get these errors, I assume that there is problem,
>>>>>>>>>>>>>> Hadoop cannot read
>>>>>>>>>>>>>> > the configuration file.
>>>>>>>>>>>>>> > I chaned the hadoop-env.sh to
>>>>>>>>>>>>>> > export
>>>>>>>>>>>>>> HADOOP_CONF_DIR=${HADOOP_CONF_DIR:-"/usr/shar/hadoop/templates/conf"}
>>>>>>>>>>>>>> > but dosen't solve the problem.
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> > ERROR org.apache.hadoop.hdfs.server.datanode.DataNode:
>>>>>>>>>>>>>> > java.lang.IllegalArgumentException: Does not contain a
>>>>>>>>>>>>>> valid host:port
>>>>>>>>>>>>>> > authority: file:/// at
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:164)
>>>>>>>>>>>>>>  at
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:201)
>>>>>>>>>>>>>> > at
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:231)
>>>>>>>>>>>>>> > at
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getServiceAddress(NameNode.java:225)
>>>>>>>>>>>>>> > at
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:347)
>>>>>>>>>>>>>> > at
>>>>>>>>>>>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.(DataNode.java:309)
>>>>>>>>>>>>>>  at
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1651)
>>>>>>>>>>>>>> > at
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1590)
>>>>>>>>>>>>>> > at
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:1608)
>>>>>>>>>>>>>> > at
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.secureMain(DataNode.java:1734)
>>>>>>>>>>>>>> > at
>>>>>>>>>>>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:1751)
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> > ________________________________
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> > FATAL org.apache.hadoop.mapred.JobTracker:
>>>>>>>>>>>>>> > java.lang.IllegalArgumentException: Does not contain a
>>>>>>>>>>>>>> valid host:port
>>>>>>>>>>>>>> > authority: local at
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:164)
>>>>>>>>>>>>>>  at
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:130)
>>>>>>>>>>>>>>  at
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> org.apache.hadoop.mapred.JobTracker.getAddress(JobTracker.java:2312)
>>>>>>>>>>>>>>  at
>>>>>>>>>>>>>> > org.apache.hadoop.mapred.JobTracker.(JobTracker.java:2070)
>>>>>>>>>>>>>> at
>>>>>>>>>>>>>> > org.apache.hadoop.mapred.JobTracker.(JobTracker.java:1889)
>>>>>>>>>>>>>> at
>>>>>>>>>>>>>> > org.apache.hadoop.mapred.JobTracker.(JobTracker.java:1883)
>>>>>>>>>>>>>> at
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> org.apache.hadoop.mapred.JobTracker.startTracker(JobTracker.java:312)
>>>>>>>>>>>>>>  at
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> org.apache.hadoop.mapred.JobTracker.startTracker(JobTracker.java:303)
>>>>>>>>>>>>>>  at
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> org.apache.hadoop.mapred.JobTracker.startTracker(JobTracker.java:298)
>>>>>>>>>>>>>> > at
>>>>>>>>>>>>>> org.apache.hadoop.mapred.JobTracker.main(JobTracker.java:4791)
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> > ________________________________
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> > ERROR org.apache.hadoop.hdfs.server.namenode.NameNode:
>>>>>>>>>>>>>> > java.lang.IllegalArgumentException: Does not contain a
>>>>>>>>>>>>>> valid host:port
>>>>>>>>>>>>>> > authority: file:/// at
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:164)
>>>>>>>>>>>>>>  at
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:201)
>>>>>>>>>>>>>> > at
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:231)
>>>>>>>>>>>>>> > at
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:265)
>>>>>>>>>>>>>> > at
>>>>>>>>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.(NameNode.java:536)
>>>>>>>>>>>>>>  at
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1410)
>>>>>>>>>>>>>> > at
>>>>>>>>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1419)
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> > ________________________________
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> > Exception in thread "main"
>>>>>>>>>>>>>> java.lang.IllegalArgumentException: Does not
>>>>>>>>>>>>>> > contain a valid host:port authority: file:/// at
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:164)
>>>>>>>>>>>>>>  at
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:201)
>>>>>>>>>>>>>> > at
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:231)
>>>>>>>>>>>>>> > at
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getServiceAddress(NameNode.java:225)
>>>>>>>>>>>>>> > at
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.initialize(SecondaryNameNode.java:167)
>>>>>>>>>>>>>> > at
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.(SecondaryNameNode.java:135)
>>>>>>>>>>>>>> > at
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.main(SecondaryNameNode.java:650)
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> > ________________________________
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> > ERROR org.apache.hadoop.mapred.TaskTracker: Can not start
>>>>>>>>>>>>>> task tracker
>>>>>>>>>>>>>> > because java.lang.IllegalArgumentException: Does not
>>>>>>>>>>>>>> contain a valid
>>>>>>>>>>>>>> > host:port authority: local at
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:164)
>>>>>>>>>>>>>>  at
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:130)
>>>>>>>>>>>>>>  at
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> org.apache.hadoop.mapred.JobTracker.getAddress(JobTracker.java:2312)
>>>>>>>>>>>>>>  at
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> org.apache.hadoop.mapred.TaskTracker.(TaskTracker.java:1532) at
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> org.apache.hadoop.mapred.TaskTracker.main(TaskTracker.java:3906)
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> > Regards,
>>>>>>>>>>>>>> > Mohammad Alkahtani
>>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>

Reply via email to