OK what the Hadoop home should be in ubuntu because the binary files in
/usr/bin
the hadoop-env.sh and othe xml file in /etc/hadoop
the conf files in /usr/share/hadoop/templates/conf

shall I use /usr as hadoop path because it is the dir that contain the bin
files

Mohammad Alkahtani
P.O.Box 102275
Riyadh 11675
Saudi Arabia
mobile: 00966 555 33 1717


On Sun, Mar 17, 2013 at 9:50 PM, Mohammad Tariq <[email protected]> wrote:

> log out from the user. log in again and see if it works.
>
> Warm Regards,
> Tariq
> https://mtariq.jux.com/
> cloudfront.blogspot.com
>
>
> On Mon, Mar 18, 2013 at 12:18 AM, Mohammad Tariq <[email protected]>wrote:
>
>> you can avoid the warning by setting the following prop to true in the
>> hadoop-env.sh file :
>> export HADOOP_HOME_WARN_SUPPRESS=true
>>
>>
>>
>> Warm Regards,
>> Tariq
>> https://mtariq.jux.com/
>> cloudfront.blogspot.com
>>
>>
>> On Mon, Mar 18, 2013 at 12:07 AM, Mohammad Alkahtani <
>> [email protected]> wrote:
>>
>>> Thank you Mohammad
>>> I still get the same error with this msg
>>>
>>> localhost: Warning: $HADOOP_HOME is deprecated.
>>> I searched ~/.bashrc but only what I wrote is there.
>>>
>>>
>>> Mohammad Alkahtani
>>> P.O.Box 102275
>>> Riyadh 11675
>>> Saudi Arabia
>>> mobile: 00966 555 33 1717
>>>
>>>
>>> On Sun, Mar 17, 2013 at 9:21 PM, Mohammad Tariq <[email protected]>wrote:
>>>
>>>> you can do that using these command :
>>>>
>>>> sudo gedit ~/.bashrc
>>>>
>>>> then go to the end of the file and add this line :
>>>> export HADOOP_HOME=/YOUR_FULL_HADOOP_PATH
>>>>
>>>> after that use it to freeze the changes :
>>>> source ~/.bashrc
>>>>
>>>> to check it :
>>>> echo $HADOOP_HOME
>>>>
>>>> This will permanently set your HADOOP_HOME.
>>>>
>>>> HTH
>>>>
>>>>
>>>> Warm Regards,
>>>> Tariq
>>>> https://mtariq.jux.com/
>>>> cloudfront.blogspot.com
>>>>
>>>>
>>>> On Sun, Mar 17, 2013 at 11:46 PM, Mohammad Alkahtani <
>>>> [email protected]> wrote:
>>>>
>>>>> Hi Tariq, Could you please tell me how to set HADOOP_HOME because I
>>>>> don't find it in the hadoop-env.sh
>>>>>
>>>>> Thank you Shashwat
>>>>> this is the output and it is already configured but hadoop don't read
>>>>> the configuration from here.
>>>>>
>>>>> /usr/share/maven-repo/org/apache
>>>>> /commons/commons-parent/22/commons-parent-22-site.xml
>>>>> /usr/share/maven-repo/org/apache/commons/commons-parent/debian
>>>>> /commons-parent-debian-site.xml
>>>>> /usr/share/maven-repo/org/apache/apache/10/apache-10-site.xml
>>>>> /usr/share/maven-repo/org/apache/apache/debian/apache-debian-site.xml
>>>>> /usr/share/compiz/composite.xml
>>>>> /usr/share/hadoop/templates/conf/mapred-site.xml
>>>>> /usr/share/hadoop/templates/conf/core-site.xml
>>>>> /usr/share/hadoop/templates/conf/hdfs-site.xml
>>>>>
>>>>> Mohammad Alkahtani
>>>>> P.O.Box 102275
>>>>> Riyadh 11675
>>>>> Saudi Arabia
>>>>> mobile: 00966 555 33 1717
>>>>>
>>>>>
>>>>> On Sun, Mar 17, 2013 at 9:07 PM, shashwat shriparv <
>>>>> [email protected]> wrote:
>>>>>
>>>>>> try
>>>>>> find / -type f -iname "*site.xml"
>>>>>> it will show you where ever those files are..
>>>>>>
>>>>>>
>>>>>>
>>>>>> ∞
>>>>>> Shashwat Shriparv
>>>>>>
>>>>>>
>>>>>>
>>>>>> On Sun, Mar 17, 2013 at 11:34 PM, Mohammad Alkahtani <
>>>>>> [email protected]> wrote:
>>>>>>
>>>>>>> The problem is I tried I read the configuration file by changing
>>>>>>> export HADOOP_CONF_DIR=${HADOOP_CONF_
>>>>>>> DIR:-"/usr/shar/hadoop/templates/conf"}
>>>>>>> but I think Hadoop dosen't get the configration from this dir, I
>>>>>>> trid and searched the system for conf dir the only dir is this one 
>>>>>>> which I
>>>>>>> changed.
>>>>>>>
>>>>>>> Mohammad Alkahtani
>>>>>>> P.O.Box 102275
>>>>>>> Riyadh 11675
>>>>>>> Saudi Arabia
>>>>>>> mobile: 00966 555 33 1717
>>>>>>>
>>>>>>>
>>>>>>> On Sun, Mar 17, 2013 at 8:57 PM, shashwat shriparv <
>>>>>>> [email protected]> wrote:
>>>>>>>
>>>>>>>> Ye its is asking for file:/// instead of hdfs:// just check if it
>>>>>>>> is taking setting configuration from other location...
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> ∞
>>>>>>>> Shashwat Shriparv
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> On Sun, Mar 17, 2013 at 11:07 PM, Luangsay Sourygna <
>>>>>>>> [email protected]> wrote:
>>>>>>>>
>>>>>>>>> Hi,
>>>>>>>>>
>>>>>>>>> What is the version of Hadoop you use?
>>>>>>>>>
>>>>>>>>> Try using fs.defaultFS instead of fs.default.name (see the list
>>>>>>>>> of all
>>>>>>>>> the deprecated properties here:
>>>>>>>>>
>>>>>>>>> http://hadoop.apache.org/docs/current/hadoop-project-dist/hadoop-common/DeprecatedProperties.html
>>>>>>>>> ).
>>>>>>>>> I remember I once had a similar error message and it was due to the
>>>>>>>>> change in properties names.
>>>>>>>>>
>>>>>>>>> Regards,
>>>>>>>>>
>>>>>>>>> Sourygna
>>>>>>>>>
>>>>>>>>> On Sun, Mar 17, 2013 at 2:32 PM, Mohammad Alkahtani
>>>>>>>>> <[email protected]> wrote:
>>>>>>>>> > Hi to all users of Hadoop,
>>>>>>>>> >
>>>>>>>>> > I installed Hadoop the .deb file on Ubuntu 12.04 but I might
>>>>>>>>> could not
>>>>>>>>> > configure it right. The conf dir is under templates in
>>>>>>>>> /usr/shar/hadoop. I
>>>>>>>>> > edit the core-site.xml, mapred-site.xml files to give
>>>>>>>>> > <property>
>>>>>>>>> > <name>fs.default.name</name>
>>>>>>>>> > <value>hdfs://localhost:9000</value>
>>>>>>>>> > </property>
>>>>>>>>> > and for mapred
>>>>>>>>> > <property>
>>>>>>>>> > <name>mapred.job.tracker</name>
>>>>>>>>> > <value>localhost:9001</value>
>>>>>>>>> > </property>
>>>>>>>>> >
>>>>>>>>> > but i get these errors, I assume that there is problem, Hadoop
>>>>>>>>> cannot read
>>>>>>>>> > the configuration file.
>>>>>>>>> > I chaned the hadoop-env.sh to
>>>>>>>>> > export
>>>>>>>>> HADOOP_CONF_DIR=${HADOOP_CONF_DIR:-"/usr/shar/hadoop/templates/conf"}
>>>>>>>>> > but dosen't solve the problem.
>>>>>>>>> >
>>>>>>>>> > ERROR org.apache.hadoop.hdfs.server.datanode.DataNode:
>>>>>>>>> > java.lang.IllegalArgumentException: Does not contain a valid
>>>>>>>>> host:port
>>>>>>>>> > authority: file:/// at
>>>>>>>>> >
>>>>>>>>> org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:164) at
>>>>>>>>> >
>>>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:201)
>>>>>>>>> > at
>>>>>>>>> >
>>>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:231)
>>>>>>>>> > at
>>>>>>>>> >
>>>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getServiceAddress(NameNode.java:225)
>>>>>>>>> > at
>>>>>>>>> >
>>>>>>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:347)
>>>>>>>>> > at
>>>>>>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.(DataNode.java:309) at
>>>>>>>>> >
>>>>>>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1651)
>>>>>>>>> > at
>>>>>>>>> >
>>>>>>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1590)
>>>>>>>>> > at
>>>>>>>>> >
>>>>>>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:1608)
>>>>>>>>> > at
>>>>>>>>> >
>>>>>>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.secureMain(DataNode.java:1734)
>>>>>>>>> > at
>>>>>>>>> org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:1751)
>>>>>>>>> >
>>>>>>>>> > ________________________________
>>>>>>>>> >
>>>>>>>>> > FATAL org.apache.hadoop.mapred.JobTracker:
>>>>>>>>> > java.lang.IllegalArgumentException: Does not contain a valid
>>>>>>>>> host:port
>>>>>>>>> > authority: local at
>>>>>>>>> >
>>>>>>>>> org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:164) at
>>>>>>>>> >
>>>>>>>>> org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:130) at
>>>>>>>>> >
>>>>>>>>> org.apache.hadoop.mapred.JobTracker.getAddress(JobTracker.java:2312) 
>>>>>>>>> at
>>>>>>>>> > org.apache.hadoop.mapred.JobTracker.(JobTracker.java:2070) at
>>>>>>>>> > org.apache.hadoop.mapred.JobTracker.(JobTracker.java:1889) at
>>>>>>>>> > org.apache.hadoop.mapred.JobTracker.(JobTracker.java:1883) at
>>>>>>>>> >
>>>>>>>>> org.apache.hadoop.mapred.JobTracker.startTracker(JobTracker.java:312) 
>>>>>>>>> at
>>>>>>>>> >
>>>>>>>>> org.apache.hadoop.mapred.JobTracker.startTracker(JobTracker.java:303) 
>>>>>>>>> at
>>>>>>>>> >
>>>>>>>>> org.apache.hadoop.mapred.JobTracker.startTracker(JobTracker.java:298)
>>>>>>>>> > at org.apache.hadoop.mapred.JobTracker.main(JobTracker.java:4791)
>>>>>>>>> >
>>>>>>>>> > ________________________________
>>>>>>>>> >
>>>>>>>>> > ERROR org.apache.hadoop.hdfs.server.namenode.NameNode:
>>>>>>>>> > java.lang.IllegalArgumentException: Does not contain a valid
>>>>>>>>> host:port
>>>>>>>>> > authority: file:/// at
>>>>>>>>> >
>>>>>>>>> org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:164) at
>>>>>>>>> >
>>>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:201)
>>>>>>>>> > at
>>>>>>>>> >
>>>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:231)
>>>>>>>>> > at
>>>>>>>>> >
>>>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:265)
>>>>>>>>> > at
>>>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.(NameNode.java:536) at
>>>>>>>>> >
>>>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1410)
>>>>>>>>> > at
>>>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1419)
>>>>>>>>> >
>>>>>>>>> > ________________________________
>>>>>>>>> >
>>>>>>>>> > Exception in thread "main" java.lang.IllegalArgumentException:
>>>>>>>>> Does not
>>>>>>>>> > contain a valid host:port authority: file:/// at
>>>>>>>>> >
>>>>>>>>> org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:164) at
>>>>>>>>> >
>>>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:201)
>>>>>>>>> > at
>>>>>>>>> >
>>>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:231)
>>>>>>>>> > at
>>>>>>>>> >
>>>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.getServiceAddress(NameNode.java:225)
>>>>>>>>> > at
>>>>>>>>> >
>>>>>>>>> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.initialize(SecondaryNameNode.java:167)
>>>>>>>>> > at
>>>>>>>>> >
>>>>>>>>> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.(SecondaryNameNode.java:135)
>>>>>>>>> > at
>>>>>>>>> >
>>>>>>>>> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.main(SecondaryNameNode.java:650)
>>>>>>>>> >
>>>>>>>>> > ________________________________
>>>>>>>>> >
>>>>>>>>> > ERROR org.apache.hadoop.mapred.TaskTracker: Can not start task
>>>>>>>>> tracker
>>>>>>>>> > because java.lang.IllegalArgumentException: Does not contain a
>>>>>>>>> valid
>>>>>>>>> > host:port authority: local at
>>>>>>>>> >
>>>>>>>>> org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:164) at
>>>>>>>>> >
>>>>>>>>> org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:130) at
>>>>>>>>> >
>>>>>>>>> org.apache.hadoop.mapred.JobTracker.getAddress(JobTracker.java:2312) 
>>>>>>>>> at
>>>>>>>>> > org.apache.hadoop.mapred.TaskTracker.(TaskTracker.java:1532) at
>>>>>>>>> > org.apache.hadoop.mapred.TaskTracker.main(TaskTracker.java:3906)
>>>>>>>>> >
>>>>>>>>> >
>>>>>>>>> > Regards,
>>>>>>>>> > Mohammad Alkahtani
>>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>

Reply via email to