Hi Df,      

      Are you use the IP instead of names in conf/masters and conf/slaves . For 
running the secondary namenode in separate machine refer the following link 



http://www.hadoop-blog.com/2010/12/secondarynamenode-process-is-starting.html 


Regards,

Shanmuganathan



---- On Tue, 16 Aug 2011 17:06:04 +0530 A 
Df<[email protected]> wrote ---- 


I already used a few tutorials as follows:
    * Hadoop Tutorial on Yahoo Developer network which uses an old hadoop and 
thus older conf files.

    * 
http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-multi-node-cluster/
 which only has two nodes and the master acts as namenode and secondary 
namenode. I need one with more than that.  


Is there a way to prevent the node from using the central file system because I 
don't have root permission and my user folder is in a central file system which 
is replicated on all the nodes?

See inline too for my responses



>________________________________
>From: Steve Loughran <[email protected]>
>To: [email protected]
>Sent: Tuesday, 16 August 2011, 12:08
>Subject: Re: hadoop cluster mode not starting up
>
>On 16/08/11 11:19, A Df wrote:
>> See inline
>>
>>
>>
>>> ________________________________
>>> From: Steve Loughran<[email protected]>
>>> To: [email protected]
>>> Sent: Tuesday, 16 August 2011, 11:08
>>> Subject: Re: hadoop cluster mode not starting up
>>>
>>> On 16/08/11 11:02, A Df wrote:
>>>> Hello All:
>>>>
>>>> I used a combination of tutorials to setup hadoop but most 
seems to be using either an old version of hadoop or only using 2 machines for 
the cluster which isn't really a cluster. Does anyone know of a good tutorial 
which setups multiple nodes for a cluster?? I already looked at the Apache 
website but it does not give sample values for the conf files. Also each set of 
tutorials seem to have a different set of parameters which they indicate should 
be changed so now its a bit confusing. For example, my configuration sets a 
dedicate namenode, secondary namenode and 8 slave nodes but when I run the 
start command it gives an error. Should I install hadoop to my user directory 
or on the root? I have it in my directory but all the nodes have a central file 
system as opposed to distributed so whatever I do on one node in my user folder 
it affect all the others so how do i set the paths to ensure that it uses a 
distributed system?
>>>>
>>>> For the errors below, I checked the directories and the files 
are there. Am I not sure what went wrong and how to set the conf to not have 
central file system. Thank you.
>>>>
>>>> Error message
>>>> CODE
>>>> w1153435@n51:~/hadoop-0.20.2_cluster>  bin/start-dfs.sh
>>>> bin/start-dfs.sh: line 28: 
/w1153435/hadoop-0.20.2_cluster/bin/hadoop-config.sh: No such file or directory
>>>> bin/start-dfs.sh: line 50: 
/w1153435/hadoop-0.20.2_cluster/bin/hadoop-daemon.sh: No such file or directory
>>>> bin/start-dfs.sh: line 51: 
/w1153435/hadoop-0.20.2_cluster/bin/hadoop-daemons.sh: No such file or directory
>>>> bin/start-dfs.sh: line 52: 
/w1153435/hadoop-0.20.2_cluster/bin/hadoop-daemons.sh: No such file or directory
>>>> CODE
>>>
>>> there's  No such file or directory as
>>> /w1153435/hadoop-0.20.2_cluster/bin/hadoop-daemons.sh
>>>
>>>
>>> There is, I checked as shown
>>> w1153435@n51:~/hadoop-0.20.2_cluster>  ls bin
>>> hadoop            rcc                start-dfs.sh      stop-dfs.sh
>>> hadoop-config.sh  slaves.sh          start-mapred.sh  
stop-mapred.sh
>>> hadoop-daemon.sh  start-all.sh      stop-all.sh
>>> hadoop-daemons.sh  start-balancer.sh  stop-balancer.sh
>
>try "pwd" to print out where the OS thinks you are, as it doesn't seem 
>to be where you think you are
>
>
>w1153435@ngs:~/hadoop-0.20.2_cluster> pwd
>/home/w1153435/hadoop-0.20.2_cluster
>
>
>w1153435@ngs:~/hadoop-0.20.2_cluster/bin> pwd
>/home/w1153435/hadoop-0.20.2_cluster/bin
>
>>>
>>>
>>>
>>>
>>>>
>>>> I had tried running this command below earlier but also got 
problems:
>>>> CODE
>>>> w1153435@ngs:~/hadoop-0.20.2_cluster>  export 
HADOOP_CONF_DIR=${HADOOP_HOME}/conf
>>>> w1153435@ngs:~/hadoop-0.20.2_cluster>  export 
HADOOP_SLAVES=${HADOOP_CONF_DIR}/slaves
>>>> w1153435@ngs:~/hadoop-0.20.2_cluster>  
${HADOOP_HOME}/bin/slaves.sh "mkdir -p 
/home/w1153435/hadoop-0.20.2_cluster/tmp/hadoop"
>>>> -bash: /bin/slaves.sh: No such file or directory
>>>> w1153435@ngs:~/hadoop-0.20.2_cluster>  export 
HADOOP_HOME=/home/w1153435/hadoop-0.20.2_cluster
>>>> w1153435@ngs:~/hadoop-0.20.2_cluster>  
${HADOOP_HOME}/bin/slaves.sh "mkdir -p 
/home/w1153435/hadoop-0.20.2_cluster/tmp/hadoop"
>>>> cat: /conf/slaves: No such file or directory
>>>> CODE
>>>>
>>> there's  No such file or directory as /conf/slaves because you set
>>> HADOOP_HOME after setting the other env variables, which are 
expanded at
>>> set-time, not run-time.
>>>
>>> I redid the command but still have errors on the slaves
>>>
>>>
>>> w1153435@n51:~/hadoop-0.20.2_cluster>  export 
HADOOP_HOME=/home/w1153435/hadoop-0.20.2_cluster
>>> w1153435@n51:~/hadoop-0.20.2_cluster>  export 
HADOOP_CONF_DIR=${HADOOP_HOME}/conf
>>> w1153435@n51:~/hadoop-0.20.2_cluster>  export 
HADOOP_SLAVES=${HADOOP_CONF_DIR}/slaves
>>> w1153435@n51:~/hadoop-0.20.2_cluster>  
${HADOOP_HOME}/bin/slaves.sh "mkdir -p 
/home/w1153435/hadoop-0.20.2_cluster/tmp/hadoop"
>>> privn51: bash: mkdir -p 
/home/w1153435/hadoop-0.20.2_cluster/tmp/hadoop: No such file or directory
>>> privn58: bash: mkdir -p 
/home/w1153435/hadoop-0.20.2_cluster/tmp/hadoop: No such file or directory
>>> privn52: bash: mkdir -p 
/home/w1153435/hadoop-0.20.2_cluster/tmp/hadoop: No such file or directory
>>> privn55: bash: mkdir -p 
/home/w1153435/hadoop-0.20.2_cluster/tmp/hadoop: No such file or directory
>>> privn57: bash: mkdir -p 
/home/w1153435/hadoop-0.20.2_cluster/tmp/hadoop: No such file or directory
>>> privn54: bash: mkdir -p 
/home/w1153435/hadoop-0.20.2_cluster/tmp/hadoop: No such file or directory
>>> privn53: bash: mkdir -p 
/home/w1153435/hadoop-0.20.2_cluster/tmp/hadoop: No such file or directory
>>> privn56: bash: mkdir -p 
/home/w1153435/hadoop-0.20.2_cluster/tmp/hadoop: No such file or directory
>
>try ssh-ing in, do it by hand, make sure you have the right permissions etc
>
>
>I reset the above path variables again and checked that they existed and 
tried the command above but same error. I used ssh with no problems and no 
password request so that is fine. What else could be wrong?
>w1153435@ngs:~/hadoop-0.20.2_cluster> echo $HADOOP_HOME                  
       /home/w1153435/hadoop-0.20.2_cluster
>w1153435@ngs:~/hadoop-0.20.2_cluster> echo $HADOOP_CONF_DIR              
       /home/w1153435/hadoop-0.20.2_cluster/conf
>w1153435@ngs:~/hadoop-0.20.2_cluster> echo $HADOOP_SLAVES                
       /home/w1153435/hadoop-0.20.2_cluster/conf/slaves
>w1153435@ngs:~/hadoop-0.20.2_cluster>
>
>
>
>
>

Reply via email to