Hi Df,
Are you use the IP instead of names in conf/masters and conf/slaves . For
running the secondary namenode in separate machine refer the following link
http://www.hadoop-blog.com/2010/12/secondarynamenode-process-is-starting.html
Regards,
Shanmuganathan
---- On Tue, 16 Aug 2011 17:06:04 +0530 A
Df<[email protected]> wrote ----
I already used a few tutorials as follows:
* Hadoop Tutorial on Yahoo Developer network which uses an old hadoop and
thus older conf files.
*
http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-multi-node-cluster/
which only has two nodes and the master acts as namenode and secondary
namenode. I need one with more than that.
Is there a way to prevent the node from using the central file system because I
don't have root permission and my user folder is in a central file system which
is replicated on all the nodes?
See inline too for my responses
>________________________________
>From: Steve Loughran <[email protected]>
>To: [email protected]
>Sent: Tuesday, 16 August 2011, 12:08
>Subject: Re: hadoop cluster mode not starting up
>
>On 16/08/11 11:19, A Df wrote:
>> See inline
>>
>>
>>
>>> ________________________________
>>> From: Steve Loughran<[email protected]>
>>> To: [email protected]
>>> Sent: Tuesday, 16 August 2011, 11:08
>>> Subject: Re: hadoop cluster mode not starting up
>>>
>>> On 16/08/11 11:02, A Df wrote:
>>>> Hello All:
>>>>
>>>> I used a combination of tutorials to setup hadoop but most
seems to be using either an old version of hadoop or only using 2 machines for
the cluster which isn't really a cluster. Does anyone know of a good tutorial
which setups multiple nodes for a cluster?? I already looked at the Apache
website but it does not give sample values for the conf files. Also each set of
tutorials seem to have a different set of parameters which they indicate should
be changed so now its a bit confusing. For example, my configuration sets a
dedicate namenode, secondary namenode and 8 slave nodes but when I run the
start command it gives an error. Should I install hadoop to my user directory
or on the root? I have it in my directory but all the nodes have a central file
system as opposed to distributed so whatever I do on one node in my user folder
it affect all the others so how do i set the paths to ensure that it uses a
distributed system?
>>>>
>>>> For the errors below, I checked the directories and the files
are there. Am I not sure what went wrong and how to set the conf to not have
central file system. Thank you.
>>>>
>>>> Error message
>>>> CODE
>>>> w1153435@n51:~/hadoop-0.20.2_cluster> bin/start-dfs.sh
>>>> bin/start-dfs.sh: line 28:
/w1153435/hadoop-0.20.2_cluster/bin/hadoop-config.sh: No such file or directory
>>>> bin/start-dfs.sh: line 50:
/w1153435/hadoop-0.20.2_cluster/bin/hadoop-daemon.sh: No such file or directory
>>>> bin/start-dfs.sh: line 51:
/w1153435/hadoop-0.20.2_cluster/bin/hadoop-daemons.sh: No such file or directory
>>>> bin/start-dfs.sh: line 52:
/w1153435/hadoop-0.20.2_cluster/bin/hadoop-daemons.sh: No such file or directory
>>>> CODE
>>>
>>> there's No such file or directory as
>>> /w1153435/hadoop-0.20.2_cluster/bin/hadoop-daemons.sh
>>>
>>>
>>> There is, I checked as shown
>>> w1153435@n51:~/hadoop-0.20.2_cluster> ls bin
>>> hadoop rcc start-dfs.sh stop-dfs.sh
>>> hadoop-config.sh slaves.sh start-mapred.sh
stop-mapred.sh
>>> hadoop-daemon.sh start-all.sh stop-all.sh
>>> hadoop-daemons.sh start-balancer.sh stop-balancer.sh
>
>try "pwd" to print out where the OS thinks you are, as it doesn't seem
>to be where you think you are
>
>
>w1153435@ngs:~/hadoop-0.20.2_cluster> pwd
>/home/w1153435/hadoop-0.20.2_cluster
>
>
>w1153435@ngs:~/hadoop-0.20.2_cluster/bin> pwd
>/home/w1153435/hadoop-0.20.2_cluster/bin
>
>>>
>>>
>>>
>>>
>>>>
>>>> I had tried running this command below earlier but also got
problems:
>>>> CODE
>>>> w1153435@ngs:~/hadoop-0.20.2_cluster> export
HADOOP_CONF_DIR=${HADOOP_HOME}/conf
>>>> w1153435@ngs:~/hadoop-0.20.2_cluster> export
HADOOP_SLAVES=${HADOOP_CONF_DIR}/slaves
>>>> w1153435@ngs:~/hadoop-0.20.2_cluster>
${HADOOP_HOME}/bin/slaves.sh "mkdir -p
/home/w1153435/hadoop-0.20.2_cluster/tmp/hadoop"
>>>> -bash: /bin/slaves.sh: No such file or directory
>>>> w1153435@ngs:~/hadoop-0.20.2_cluster> export
HADOOP_HOME=/home/w1153435/hadoop-0.20.2_cluster
>>>> w1153435@ngs:~/hadoop-0.20.2_cluster>
${HADOOP_HOME}/bin/slaves.sh "mkdir -p
/home/w1153435/hadoop-0.20.2_cluster/tmp/hadoop"
>>>> cat: /conf/slaves: No such file or directory
>>>> CODE
>>>>
>>> there's No such file or directory as /conf/slaves because you set
>>> HADOOP_HOME after setting the other env variables, which are
expanded at
>>> set-time, not run-time.
>>>
>>> I redid the command but still have errors on the slaves
>>>
>>>
>>> w1153435@n51:~/hadoop-0.20.2_cluster> export
HADOOP_HOME=/home/w1153435/hadoop-0.20.2_cluster
>>> w1153435@n51:~/hadoop-0.20.2_cluster> export
HADOOP_CONF_DIR=${HADOOP_HOME}/conf
>>> w1153435@n51:~/hadoop-0.20.2_cluster> export
HADOOP_SLAVES=${HADOOP_CONF_DIR}/slaves
>>> w1153435@n51:~/hadoop-0.20.2_cluster>
${HADOOP_HOME}/bin/slaves.sh "mkdir -p
/home/w1153435/hadoop-0.20.2_cluster/tmp/hadoop"
>>> privn51: bash: mkdir -p
/home/w1153435/hadoop-0.20.2_cluster/tmp/hadoop: No such file or directory
>>> privn58: bash: mkdir -p
/home/w1153435/hadoop-0.20.2_cluster/tmp/hadoop: No such file or directory
>>> privn52: bash: mkdir -p
/home/w1153435/hadoop-0.20.2_cluster/tmp/hadoop: No such file or directory
>>> privn55: bash: mkdir -p
/home/w1153435/hadoop-0.20.2_cluster/tmp/hadoop: No such file or directory
>>> privn57: bash: mkdir -p
/home/w1153435/hadoop-0.20.2_cluster/tmp/hadoop: No such file or directory
>>> privn54: bash: mkdir -p
/home/w1153435/hadoop-0.20.2_cluster/tmp/hadoop: No such file or directory
>>> privn53: bash: mkdir -p
/home/w1153435/hadoop-0.20.2_cluster/tmp/hadoop: No such file or directory
>>> privn56: bash: mkdir -p
/home/w1153435/hadoop-0.20.2_cluster/tmp/hadoop: No such file or directory
>
>try ssh-ing in, do it by hand, make sure you have the right permissions etc
>
>
>I reset the above path variables again and checked that they existed and
tried the command above but same error. I used ssh with no problems and no
password request so that is fine. What else could be wrong?
>w1153435@ngs:~/hadoop-0.20.2_cluster> echo $HADOOP_HOME
/home/w1153435/hadoop-0.20.2_cluster
>w1153435@ngs:~/hadoop-0.20.2_cluster> echo $HADOOP_CONF_DIR
/home/w1153435/hadoop-0.20.2_cluster/conf
>w1153435@ngs:~/hadoop-0.20.2_cluster> echo $HADOOP_SLAVES
/home/w1153435/hadoop-0.20.2_cluster/conf/slaves
>w1153435@ngs:~/hadoop-0.20.2_cluster>
>
>
>
>
>