I already used a few tutorials as follows:
* Hadoop Tutorial on Yahoo Developer network which uses an old hadoop
and thus older conf files.
*
http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-multi-node-cluster/ which
only has two nodes and the master acts as namenode and secondary namenode. I
need one with more than that.
Is there a way to prevent the node from using the central file system because I
don't have root permission and my user folder is in a central file system which
is replicated on all the nodes?
See inline too for my responses
>________________________________
>From: Steve Loughran <[email protected]>
>To: [email protected]
>Sent: Tuesday, 16 August 2011, 12:08
>Subject: Re: hadoop cluster mode not starting up
>
>On 16/08/11 11:19, A Df wrote:
>> See inline
>>
>>
>>
>>> ________________________________
>>> From: Steve Loughran<[email protected]>
>>> To: [email protected]
>>> Sent: Tuesday, 16 August 2011, 11:08
>>> Subject: Re: hadoop cluster mode not starting up
>>>
>>> On 16/08/11 11:02, A Df wrote:
>>>> Hello All:
>>>>
>>>> I used a combination of tutorials to setup hadoop but most seems to be
>>>> using either an old version of hadoop or only using 2 machines for the
>>>> cluster which isn't really a cluster. Does anyone know of a good tutorial
>>>> which setups multiple nodes for a cluster?? I already looked at the Apache
>>>> website but it does not give sample values for the conf files. Also each
>>>> set of tutorials seem to have a different set of parameters which they
>>>> indicate should be changed so now its a bit confusing. For example, my
>>>> configuration sets a dedicate namenode, secondary namenode and 8 slave
>>>> nodes but when I run the start command it gives an error. Should I install
>>>> hadoop to my user directory or on the root? I have it in my directory but
>>>> all the nodes have a central file system as opposed to distributed so
>>>> whatever I do on one node in my user folder it affect all the others so
>>>> how do i set the paths to ensure that it uses a distributed system?
>>>>
>>>> For the errors below, I checked the directories and the files are there.
>>>> Am I not sure what went wrong and how to set the conf to not have central
>>>> file system. Thank you.
>>>>
>>>> Error message
>>>> CODE
>>>> w1153435@n51:~/hadoop-0.20.2_cluster> bin/start-dfs.sh
>>>> bin/start-dfs.sh: line 28:
>>>> /w1153435/hadoop-0.20.2_cluster/bin/hadoop-config.sh: No such file or
>>>> directory
>>>> bin/start-dfs.sh: line 50:
>>>> /w1153435/hadoop-0.20.2_cluster/bin/hadoop-daemon.sh: No such file or
>>>> directory
>>>> bin/start-dfs.sh: line 51:
>>>> /w1153435/hadoop-0.20.2_cluster/bin/hadoop-daemons.sh: No such file or
>>>> directory
>>>> bin/start-dfs.sh: line 52:
>>>> /w1153435/hadoop-0.20.2_cluster/bin/hadoop-daemons.sh: No such file or
>>>> directory
>>>> CODE
>>>
>>> there's No such file or directory as
>>> /w1153435/hadoop-0.20.2_cluster/bin/hadoop-daemons.sh
>>>
>>>
>>> There is, I checked as shown
>>> w1153435@n51:~/hadoop-0.20.2_cluster> ls bin
>>> hadoop rcc start-dfs.sh stop-dfs.sh
>>> hadoop-config.sh slaves.sh start-mapred.sh stop-mapred.sh
>>> hadoop-daemon.sh start-all.sh stop-all.sh
>>> hadoop-daemons.sh start-balancer.sh stop-balancer.sh
>
>try "pwd" to print out where the OS thinks you are, as it doesn't seem
>to be where you think you are
>
>
>w1153435@ngs:~/hadoop-0.20.2_cluster> pwd
>/home/w1153435/hadoop-0.20.2_cluster
>
>
>w1153435@ngs:~/hadoop-0.20.2_cluster/bin> pwd
>/home/w1153435/hadoop-0.20.2_cluster/bin
>
>>>
>>>
>>>
>>>
>>>>
>>>> I had tried running this command below earlier but also got problems:
>>>> CODE
>>>> w1153435@ngs:~/hadoop-0.20.2_cluster> export
>>>> HADOOP_CONF_DIR=${HADOOP_HOME}/conf
>>>> w1153435@ngs:~/hadoop-0.20.2_cluster> export
>>>> HADOOP_SLAVES=${HADOOP_CONF_DIR}/slaves
>>>> w1153435@ngs:~/hadoop-0.20.2_cluster> ${HADOOP_HOME}/bin/slaves.sh
>>>> "mkdir -p /home/w1153435/hadoop-0.20.2_cluster/tmp/hadoop"
>>>> -bash: /bin/slaves.sh: No such file or directory
>>>> w1153435@ngs:~/hadoop-0.20.2_cluster> export
>>>> HADOOP_HOME=/home/w1153435/hadoop-0.20.2_cluster
>>>> w1153435@ngs:~/hadoop-0.20.2_cluster> ${HADOOP_HOME}/bin/slaves.sh
>>>> "mkdir -p /home/w1153435/hadoop-0.20.2_cluster/tmp/hadoop"
>>>> cat: /conf/slaves: No such file or directory
>>>> CODE
>>>>
>>> there's No such file or directory as /conf/slaves because you set
>>> HADOOP_HOME after setting the other env variables, which are expanded at
>>> set-time, not run-time.
>>>
>>> I redid the command but still have errors on the slaves
>>>
>>>
>>> w1153435@n51:~/hadoop-0.20.2_cluster> export
>>> HADOOP_HOME=/home/w1153435/hadoop-0.20.2_cluster
>>> w1153435@n51:~/hadoop-0.20.2_cluster> export
>>> HADOOP_CONF_DIR=${HADOOP_HOME}/conf
>>> w1153435@n51:~/hadoop-0.20.2_cluster> export
>>> HADOOP_SLAVES=${HADOOP_CONF_DIR}/slaves
>>> w1153435@n51:~/hadoop-0.20.2_cluster> ${HADOOP_HOME}/bin/slaves.sh "mkdir
>>> -p /home/w1153435/hadoop-0.20.2_cluster/tmp/hadoop"
>>> privn51: bash: mkdir -p /home/w1153435/hadoop-0.20.2_cluster/tmp/hadoop: No
>>> such file or directory
>>> privn58: bash: mkdir -p /home/w1153435/hadoop-0.20.2_cluster/tmp/hadoop: No
>>> such file or directory
>>> privn52: bash: mkdir -p /home/w1153435/hadoop-0.20.2_cluster/tmp/hadoop: No
>>> such file or directory
>>> privn55: bash: mkdir -p /home/w1153435/hadoop-0.20.2_cluster/tmp/hadoop: No
>>> such file or directory
>>> privn57: bash: mkdir -p /home/w1153435/hadoop-0.20.2_cluster/tmp/hadoop: No
>>> such file or directory
>>> privn54: bash: mkdir -p /home/w1153435/hadoop-0.20.2_cluster/tmp/hadoop: No
>>> such file or directory
>>> privn53: bash: mkdir -p /home/w1153435/hadoop-0.20.2_cluster/tmp/hadoop: No
>>> such file or directory
>>> privn56: bash: mkdir -p /home/w1153435/hadoop-0.20.2_cluster/tmp/hadoop: No
>>> such file or directory
>
>try ssh-ing in, do it by hand, make sure you have the right permissions etc
>
>
>I reset the above path variables again and checked that they existed and tried
>the command above but same error. I used ssh with no problems and no password
>request so that is fine. What else could be wrong?
>w1153435@ngs:~/hadoop-0.20.2_cluster> echo $HADOOP_HOME
> /home/w1153435/hadoop-0.20.2_cluster
>w1153435@ngs:~/hadoop-0.20.2_cluster> echo $HADOOP_CONF_DIR
> /home/w1153435/hadoop-0.20.2_cluster/conf
>w1153435@ngs:~/hadoop-0.20.2_cluster> echo $HADOOP_SLAVES
> /home/w1153435/hadoop-0.20.2_cluster/conf/slaves
>w1153435@ngs:~/hadoop-0.20.2_cluster>
>
>
>
>
>