Yes - not using /dev/mapper/hdvg-rootlv was what I was planning to suggest. 

It seems to me that you installed mod_passenger and/or ambari on vbaby1. Is 
this your new ambari master? 

Try doing this on vbaby1: 

$puppet master --no-daemonize --debug

The above will create the cert required by the puppet master running in httpd. 
Kill the above process ( it will run in the foreground ). 

Now, try the httpd restart. 

 ( Also, note that you should not need to do anything for ganglia start unless 
you are running ganglia server on the same host as ambari. )

thanks
-- Hitesh


On Aug 19, 2012, at 6:03 AM, xu peng wrote:

> Hi Hitesh :
> 
> It is me again.
> 
> I figured out the previous problem by changing the mount point to a custom 
> path.
> 
> But i failed at the step of starting ganglia server.
> 
> I run this command manualy on vbaby1 node , but failed . The other
> node successed.
> ([root@vbaby1 log]# service httpd start
> Starting httpd: Syntax error on line 37 of 
> /etc/httpd/conf.d/puppetmaster.conf:
> SSLCertificateChainFile: file '/var/lib/puppet/ssl/ca/ca_crt.pem' does
> not exist or is empty
>                                                           [FAILED])
> 
> Please refer to the error log.
> 
> Thanks a lot.
> 
> On Sun, Aug 19, 2012 at 7:26 PM, xu peng <[email protected]> wrote:
>> Hi Hitesh :
>> 
>> It is me again.
>> 
>> I encountered another problem while deploying the service. And
>> according to the log , it seems like something went wrong when
>> executing command (Dependency Exec[mkdir -p
>> /dev/mapper/hdvg-rootlv/hadoop/hdfs/data] has failures: true) .
>> 
>> Please refer to the attachment. It seems like all the rpm package
>> installed successfully , and i don't know where failed the dependency.
>> 
>> Please help , thanks a lot.
>> 
>> On Sun, Aug 19, 2012 at 8:08 AM, xu peng <[email protected]> wrote:
>>> Hi Hitesh :
>>> 
>>> I use the default settings of the mount point , but it seems like this
>>> path is not a directory(/dev/mapper/hdvg-rootlv/), and i can not
>>> execute mkdir -p command on this path. And the hdvg-rootlv is a
>>> blocking file (bwrxwrxwrx) .  Is there something wrong ?
>>> 
>>> 
>>> 
>>> On Sun, Aug 19, 2012 at 3:38 AM, Hitesh Shah <[email protected]> wrote:
>>>> Hi
>>>> 
>>>> Yes - you should all packages from the new repo and none from the old 
>>>> repo. Most of the packages should be the same but same like hadoop-lzo 
>>>> were re-factored to work correctly with respect to 32/64-bit installs on 
>>>> RHEL6.
>>>> 
>>>> Regarding the mount points, from a hadoop point of view, the namenode and 
>>>> datanode dirs are just dirs. From a performance point of view, you want 
>>>> each dir to be created on a separate mount point to increase disk io 
>>>> bandwidth. This means that the mount points that you select on the UI 
>>>> should allow directories to be created. If you have mounted certain kind 
>>>> of filesystems which you do not wish to use for hadoop ( any tmpfs, nfs 
>>>> mounts etc ), you should de-select them on the UI and/or use the custom 
>>>> mount point text box as appropriate. The UI currently does not distinguish 
>>>> valid mount points and therefore it is up to the user to select correctly.
>>>> 
>>>> -- Hitesh
>>>> 
>>>> 
>>>> On Aug 18, 2012, at 9:48 AM, xu peng wrote:
>>>> 
>>>>> Hi Hitesh:
>>>>> 
>>>>> Thanks again for your reply.
>>>>> 
>>>>> I solved the dependency problem after updating the hdp repo.
>>>>> 
>>>>> But here comes two new problems :
>>>>> 1. I update the new hdp repo , but i create a local repo copy of the
>>>>> old hdp repo. And I installed all the rpm package except
>>>>> hadoop-lzo-native using the old hdp repo. So it seems like the
>>>>> hadoop-lzo-native has some conflct with hadoop-lzo. So , do i have to
>>>>> install all the rpm package from the new repo ?
>>>>> 
>>>>> 2. From the error log , i can see a command "mkdir -p /var/.../..
>>>>> (mounting point of hadoop)", but i found the mouting point is not a
>>>>> dir , but a blocking file(bwrxwrxwrx). And the execution of this step
>>>>> failed. Did i do something wrong ?
>>>>> 
>>>>> I am sorry that this deploy error log is on my company's computer, and
>>>>> i will upload it in my next email.
>>>>> 
>>>>> 
>>>>> Thanks
>>>>> -- Xupeng
>>>>> 
>>>>> On Sat, Aug 18, 2012 at 4:43 AM, Hitesh Shah <[email protected]> 
>>>>> wrote:
>>>>>> Hi again,
>>>>>> 
>>>>>> You are actually hitting a problem caused by some changes in the code 
>>>>>> which require a modified repo. Unfortunately, I got delayed in modifying 
>>>>>> the documentation to point to the new repo.
>>>>>> 
>>>>>> Could you try using
>>>>>> http://public-repo-1.hortonworks.com/HDP-1.0.1.14/repos/centos5/hdp-release-1.0.1.14-1.el5.noarch.rpm
>>>>>> or
>>>>>> http://public-repo-1.hortonworks.com/HDP-1.0.1.14/repos/centos6/hdp-release-1.0.1.14-1.el6.noarch.rpm
>>>>>> 
>>>>>> The above should install the yum repo configs to point to the correct 
>>>>>> repo which will have the lzo packages.
>>>>>> 
>>>>>> -- Hitesh
>>>>>> 
>>>>>> 
>>>>>> On Aug 16, 2012, at 9:27 PM, xu peng wrote:
>>>>>> 
>>>>>>> Hitesh Shah :
>>>>>>> 
>>>>>>> It is a my my pleasure to fill jira of ambari to help other users . As
>>>>>>> a matter of fact, i want to summarize all the problem before i install
>>>>>>> ambari cluster successfully. And i will  feed back as soon as
>>>>>>> possiable.
>>>>>>> 
>>>>>>> Here is another problem i encounter when install hadoop using ambari,
>>>>>>> i found a rpm package "hadoop-lzp-native" not in the hdp repo
>>>>>>> (baseurl=http://public-repo-1.hortonworks.com/HDP-1.0.13/repos/centos5)
>>>>>>> . So i failed againg during deploying step.
>>>>>>> 
>>>>>>> And the attachment is the deploying log , please refer.
>>>>>>> 
>>>>>>> Thanks a lot and look forward to you reply.
>>>>>>> 
>>>>>>> 
>>>>>>> On Tue, Aug 14, 2012 at 11:35 PM, Hitesh Shah <[email protected]> 
>>>>>>> wrote:
>>>>>>>> Ok - the cert issue is sometimes a result of uninstalling and 
>>>>>>>> re-installing ambari agents.
>>>>>>>> 
>>>>>>>> The re-install causes ambari agents to regenerate a new certification 
>>>>>>>> and if the master was bootstrapped earlier, it would still be looking 
>>>>>>>> to match against old certs.
>>>>>>>> 
>>>>>>>> Stop ambari master and remove ambari-agent rpm from all hosts.
>>>>>>>> 
>>>>>>>> To fix this:
>>>>>>>> - on the master, do a puppet cert revoke for all hosts ( 
>>>>>>>> http://docs.puppetlabs.com/man/cert.html )
>>>>>>>> - you can do a cert list to get all signed or non-signed hosts
>>>>>>>> 
>>>>>>>> On all hosts, delete the following dirs ( if they exist ) :
>>>>>>>> - /etc/puppet/ssl
>>>>>>>> - /etc/puppet/[master|agent\/ssl/
>>>>>>>> - /var/lib/puppet/ssl/
>>>>>>>> 
>>>>>>>> 
>>>>>>>> After doing the above, re-install the ambari agent.
>>>>>>>> 
>>>>>>>> On the ambari master, stop the master. Run the following command:
>>>>>>>> 
>>>>>>>> puppet master --no-daemonize --debug
>>>>>>>> 
>>>>>>>> The above runs in the foreground. The reason to run this is to make 
>>>>>>>> sure the cert for the master is recreated as we deleted it earlier.
>>>>>>>> 
>>>>>>>> Now, kill the above process running in the foreground and do a service 
>>>>>>>> ambari start to bring up the UI.
>>>>>>>> 
>>>>>>>> You should be able to bootstrap from this point on.
>>>>>>>> 
>>>>>>>> Would you mind filing a jira and mentioning all the various issues you 
>>>>>>>> have come across and how you solved them. We can use that to create an 
>>>>>>>> FAQ for other users.
>>>>>>>> 
>>>>>>>> thanks
>>>>>>>> -- Hitesh
>>>>>>>> 
>>>>>>>> 
>>>>>>>> On Aug 14, 2012, at 1:55 AM, xu peng wrote:
>>>>>>>> 
>>>>>>>>> Hi  Hitesh :
>>>>>>>>> 
>>>>>>>>> Thanks a lot for your reply.
>>>>>>>>> 
>>>>>>>>> 1. I did a puppet kick --ping to the client from my ambari master ,
>>>>>>>>> all the five nodes failed with the same log (Triggering
>>>>>>>>> vbaby2.cloud.eb
>>>>>>>>> Host vbaby2.cloud.eb failed: certificate verify failed.  This is often
>>>>>>>>> because the time is out of sync on the server or client
>>>>>>>>> vbaby2.cloud.eb finished with exit code 2)
>>>>>>>>> 
>>>>>>>>> I manually run "service ambari-agent start" , is that necessary ? How
>>>>>>>>> can i fix these problem ?
>>>>>>>>> 
>>>>>>>>> 2. As you suggest , I run the yum command manually. And found that the
>>>>>>>>> installation missed some dependecy - php-gd. And i have to update my
>>>>>>>>> yum repo.
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> On Tue, Aug 14, 2012 at 1:01 AM, Hitesh Shah <[email protected]> 
>>>>>>>>> wrote:
>>>>>>>>>> Based on your deploy error log:
>>>>>>>>>> 
>>>>>>>>>> "3": {
>>>>>>>>>>     "nodeReport": {
>>>>>>>>>>         "PUPPET_KICK_FAILED": [],
>>>>>>>>>>         "PUPPET_OPERATION_FAILED": [
>>>>>>>>>>             "vbaby3.cloud.eb",
>>>>>>>>>>             "vbaby5.cloud.eb",
>>>>>>>>>>             "vbaby4.cloud.eb",
>>>>>>>>>>             "vbaby2.cloud.eb",
>>>>>>>>>>             "vbaby6.cloud.eb",
>>>>>>>>>>             "vbaby1.cloud.eb"
>>>>>>>>>>         ],
>>>>>>>>>>         "PUPPET_OPERATION_TIMEDOUT": [
>>>>>>>>>>             "vbaby5.cloud.eb",
>>>>>>>>>>             "vbaby4.cloud.eb",
>>>>>>>>>>             "vbaby2.cloud.eb",
>>>>>>>>>>             "vbaby6.cloud.eb",
>>>>>>>>>>             "vbaby1.cloud.eb"
>>>>>>>>>>         ],
>>>>>>>>>> 
>>>>>>>>>> 5 nodes timed out which means the puppet agent is not running on 
>>>>>>>>>> them or they cannot communicate with the master. Trying doing a 
>>>>>>>>>> puppet kick --ping to them from the master.
>>>>>>>>>> 
>>>>>>>>>> For the one which failed, it failed at
>>>>>>>>>> 
>>>>>>>>>> "\"Mon Aug 13 11:54:17 +0800 2012 
>>>>>>>>>> /Stage[1]/Hdp::Pre_install_pkgs/Hdp::Exec[yum install 
>>>>>>>>>> $pre_installed_pkgs]/Exec[yum install $pre_installed_pkgs]/returns 
>>>>>>>>>> (err): change from notrun to 0 failed: yum install -y hadoop 
>>>>>>>>>> hadoop-libhdfs hadoop-native hadoop-pipes hadoop-sbin hadoop-lzo 
>>>>>>>>>> hadoop hadoop-libhdfs hadoop-native hadoop-pipes hadoop-sbin 
>>>>>>>>>> hadoop-lzo hdp_mon_dashboard ganglia-gmond-3.2.0 gweb 
>>>>>>>>>> hdp_mon_ganglia_addons snappy snappy-devel returned 1 instead of one 
>>>>>>>>>> of [0] at /etc/puppet/agent/modules/hdp/manifests/init.pp:265\"",
>>>>>>>>>> 
>>>>>>>>>> It seems like yum install failed on the host. Try running the 
>>>>>>>>>> command manually and see what the error is.
>>>>>>>>>> 
>>>>>>>>>> -- Hitesh
>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>>> On Aug 13, 2012, at 2:28 AM, xu peng wrote:
>>>>>>>>>> 
>>>>>>>>>>> Hi Hitesh :
>>>>>>>>>>> 
>>>>>>>>>>> It's me again.
>>>>>>>>>>> 
>>>>>>>>>>> Followed you advice , I reinstalled the ambari server. But deploying
>>>>>>>>>>> cluster and uninstall cluster failed again. I really  don't know 
>>>>>>>>>>> why.
>>>>>>>>>>> 
>>>>>>>>>>> I supplied a attachment which contains the logs of  all the nodes in
>>>>>>>>>>> my cluster (/var/log/puppet_*.log , /var/log/puppet/*.log ,
>>>>>>>>>>> /var/log/yum.log, /var/log/hmc/hmc.log). And vbaby3.cloud.eb is the
>>>>>>>>>>> ambari server. Please refer.
>>>>>>>>>>> 
>>>>>>>>>>> Attachment DeployError and UninstallError is the log supplied by the
>>>>>>>>>>> website of ambari when failing. And attachment DeployingDetails.jpg 
>>>>>>>>>>> is
>>>>>>>>>>> the deploy details of my cluster. Please refer.
>>>>>>>>>>> 
>>>>>>>>>>> 
>>>>>>>>>>> Thanks again for your patience ! And look forward to your reply.
>>>>>>>>>>> 
>>>>>>>>>>> Xupeng
>>>>>>>>>>> 
>>>>>>>>>>> On Sat, Aug 11, 2012 at 10:56 PM, Hitesh Shah 
>>>>>>>>>>> <[email protected]> wrote:
>>>>>>>>>>>> For uninstall failures, you will need to do a couple of things. 
>>>>>>>>>>>> Depending on where the uninstall failed, you may have to manually 
>>>>>>>>>>>> do a killall java on all the nodes to kill any missed processes. 
>>>>>>>>>>>> If you want to start with a complete clean install, you should 
>>>>>>>>>>>> also delete the hadoop dir in the mount points you selected during 
>>>>>>>>>>>> the previous install  so that the new fresh install does not face 
>>>>>>>>>>>> errors when it tries to re-format hdfs.
>>>>>>>>>>>> 
>>>>>>>>>>>> After that, simply, uinstall and re-install ambari rpm and that 
>>>>>>>>>>>> should allow you to re-create a fresh cluster.
>>>>>>>>>>>> 
>>>>>>>>>>>> -- Hitesh
>>>>>>>>>>>> 
>>>>>>>>>>>> On Aug 11, 2012, at 2:34 AM, xu peng wrote:
>>>>>>>>>>>> 
>>>>>>>>>>>>> Hi Hitesh :
>>>>>>>>>>>>> 
>>>>>>>>>>>>> Thanks a lot for your reply.
>>>>>>>>>>>>> 
>>>>>>>>>>>>> I solved this problem , it is silly mistake. Someone has changed 
>>>>>>>>>>>>> the
>>>>>>>>>>>>> owner of "/" dir , and according to the errorlog , pdsh need root 
>>>>>>>>>>>>> to
>>>>>>>>>>>>> proceed.
>>>>>>>>>>>>> 
>>>>>>>>>>>>> After changing the owner of "/" to root , problem solved. Thank 
>>>>>>>>>>>>> you
>>>>>>>>>>>>> again for you reply.
>>>>>>>>>>>>> 
>>>>>>>>>>>>> I have another question. I had a uninstall failure , and there is 
>>>>>>>>>>>>> no
>>>>>>>>>>>>> button on the website for me to rollback and i don't know what to 
>>>>>>>>>>>>> do
>>>>>>>>>>>>> about that. What should i do now to reinstall hadoop ?
>>>>>>>>>>>>> 
>>>>>>>>>>>>> Thanks
>>>>>>>>>>>>> 
>>>>>>>>>>>>> On Fri, Aug 10, 2012 at 10:55 PM, Hitesh Shah 
>>>>>>>>>>>>> <[email protected]> wrote:
>>>>>>>>>>>>>> Hi
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> Currently, the ambari installer requires everything to be run as 
>>>>>>>>>>>>>> root. It does not detect that the user is not root and use sudo 
>>>>>>>>>>>>>> either on the master or on the agent nodes.
>>>>>>>>>>>>>> Furthermore, it seems like it is failing when trying to use pdsh 
>>>>>>>>>>>>>> to make remote calls to the host list that you passed in due to 
>>>>>>>>>>>>>> the errors mentioned in your script. This could be due to how it 
>>>>>>>>>>>>>> was installed but I am not sure.
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> Could you switch to become root and run any simple command on 
>>>>>>>>>>>>>> all hosts using pdsh? If you want to reference exactly how 
>>>>>>>>>>>>>> ambari uses pdsh, you can look into 
>>>>>>>>>>>>>> /usr/share/hmc/php/frontend/commandUtils.php
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> thanks
>>>>>>>>>>>>>> -- Hitesh
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> On Aug 9, 2012, at 9:04 PM, xu peng wrote:
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> According to the error log , is there something wrong with my 
>>>>>>>>>>>>>>> account ?
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> I installed all the dependency module and ambari with the user
>>>>>>>>>>>>>>> "ambari" instead of root. I added user "ambari" to 
>>>>>>>>>>>>>>> /etc/sudofilers
>>>>>>>>>>>>>>> with no passwd.
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> On Fri, Aug 10, 2012 at 11:49 AM, xu peng 
>>>>>>>>>>>>>>> <[email protected]> wrote:
>>>>>>>>>>>>>>>> There is no 100.log.file in /var/log/hmc dir, but only 55.log 
>>>>>>>>>>>>>>>> file (55
>>>>>>>>>>>>>>>> is the biggest version num).
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> The content of 55.log is :
>>>>>>>>>>>>>>>> pdsh@vbaby1: module path "/usr/lib64/pdsh" insecure.
>>>>>>>>>>>>>>>> pdsh@vbaby1: "/": Owner not root, current uid, or pdsh 
>>>>>>>>>>>>>>>> executable owner
>>>>>>>>>>>>>>>> pdsh@vbaby1: Couldn't load any pdsh modules
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> Thanks ~
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> On Fri, Aug 10, 2012 at 11:36 AM, Hitesh Shah 
>>>>>>>>>>>>>>>> <[email protected]> wrote:
>>>>>>>>>>>>>>>>> Sorry - my mistake. The last txn mentioned is 100 so please 
>>>>>>>>>>>>>>>>> look for the 100.log file.
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> -- Hitesh
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> On Aug 9, 2012, at 8:34 PM, Hitesh Shah wrote:
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>> Thanks - will take a look and get back to you.
>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>> Could you also look at /var/log/hmc/hmc.txn.55.log and see 
>>>>>>>>>>>>>>>>>> if there are any errors in it?
>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>> -- Hitesh.
>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>> On Aug 9, 2012, at 8:00 PM, xu peng wrote:
>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> Hi Hitesh :
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> Thanks a lot for your replying. I have done all your 
>>>>>>>>>>>>>>>>>>> suggestions in my
>>>>>>>>>>>>>>>>>>> ambari server , and the result is as below.
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> 1. I can confirm that the hosts.txt file is empty after i 
>>>>>>>>>>>>>>>>>>> failed at
>>>>>>>>>>>>>>>>>>> the step finding reachable nodes.
>>>>>>>>>>>>>>>>>>> 2. I tried make hostdetails file in win7 and redhat , it 
>>>>>>>>>>>>>>>>>>> both
>>>>>>>>>>>>>>>>>>> failed.(Please see the attachment, my hostdetails file)
>>>>>>>>>>>>>>>>>>> 3. I removed the logging re-direct and run the .sh script 
>>>>>>>>>>>>>>>>>>> .It seems
>>>>>>>>>>>>>>>>>>> like the script works well , it print the hostname in 
>>>>>>>>>>>>>>>>>>> console and
>>>>>>>>>>>>>>>>>>> generate a file (content  is "0") in the same dir. (Please 
>>>>>>>>>>>>>>>>>>> see the
>>>>>>>>>>>>>>>>>>> attachment , the result and my .sh script )
>>>>>>>>>>>>>>>>>>> 4. I attached the hmc.log and error_log too. Hope this 
>>>>>>>>>>>>>>>>>>> helps ~
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> Thanks ~
>>>>>>>>>>>>>>>>>>> Xupeng
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> On Fri, Aug 10, 2012 at 12:24 AM, Hitesh Shah 
>>>>>>>>>>>>>>>>>>> <[email protected]> wrote:
>>>>>>>>>>>>>>>>>>>> Xupeng, can you confirm that the hosts.txt file at 
>>>>>>>>>>>>>>>>>>>> /var/run/hmc/clusters/EBHadoop/hosts.txt is empty?
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>> Also, can you ensure that the hostdetails file that you 
>>>>>>>>>>>>>>>>>>>> upload does not have any special characters that may be 
>>>>>>>>>>>>>>>>>>>> creating problems for the parsing layer?
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>> In the same dir, there should be an ssh.sh script. Can you 
>>>>>>>>>>>>>>>>>>>> create a copy of it, edit to remove the logging re-directs 
>>>>>>>>>>>>>>>>>>>> to files and run the script manually from command-line ( 
>>>>>>>>>>>>>>>>>>>> it takes in a hostname as the argument ) ? The output of 
>>>>>>>>>>>>>>>>>>>> that should show you as to what is going wrong.
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>> Also, please look at /var/log/hmc/hmc.log and 
>>>>>>>>>>>>>>>>>>>> httpd/error_log to see if there are any errors being 
>>>>>>>>>>>>>>>>>>>> logged which may shed more light on the issue.
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>> thanks
>>>>>>>>>>>>>>>>>>>> -- Hitesh
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>> On Aug 9, 2012, at 9:11 AM, Artem Ervits wrote:
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>>> Which file are you supplying in the step? Hostdetail.txt 
>>>>>>>>>>>>>>>>>>>>> or hosts?
>>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>>> From: xupeng.bupt [mailto:[email protected]]
>>>>>>>>>>>>>>>>>>>>> Sent: Thursday, August 09, 2012 11:33 AM
>>>>>>>>>>>>>>>>>>>>> To: ambari-user
>>>>>>>>>>>>>>>>>>>>> Subject: Re: RE: Problem when setting up hadoop cluster 
>>>>>>>>>>>>>>>>>>>>> step 2
>>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>>> Thank you for your replying ~
>>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>>> I made only one hostdetail.txt file which contains the 
>>>>>>>>>>>>>>>>>>>>> names of all servers. And i submit this file on the 
>>>>>>>>>>>>>>>>>>>>> website ,  but i still have the same problem. I failed at 
>>>>>>>>>>>>>>>>>>>>> the step of finding reachable nodes.
>>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>>> The error log is : "
>>>>>>>>>>>>>>>>>>>>> [ERROR][sequentialScriptExecutor][sequentialScriptRunner.php:272][]:
>>>>>>>>>>>>>>>>>>>>> Encountered total failure in transaction 100 while 
>>>>>>>>>>>>>>>>>>>>> running cmd:
>>>>>>>>>>>>>>>>>>>>> /usr/bin/php ./addNodes/findSshableNodes.php with args: 
>>>>>>>>>>>>>>>>>>>>> EBHadoop root
>>>>>>>>>>>>>>>>>>>>> 35 100 36 /var/run/hmc/clusters/EBHadoop/hosts.txt
>>>>>>>>>>>>>>>>>>>>> "
>>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>>> And my hostdetail.txt file is :"
>>>>>>>>>>>>>>>>>>>>> vbaby2.cloud.eb
>>>>>>>>>>>>>>>>>>>>> vbaby3.cloud.eb
>>>>>>>>>>>>>>>>>>>>> vbaby4.cloud.eb
>>>>>>>>>>>>>>>>>>>>> vbaby5.cloud.eb
>>>>>>>>>>>>>>>>>>>>> vbaby6.cloud.eb
>>>>>>>>>>>>>>>>>>>>> "
>>>>>>>>>>>>>>>>>>>>> Thank you very much ~
>>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>>> 2012-08-09
>>>>>>>>>>>>>>>>>>>>> xupeng.bupt
>>>>>>>>>>>>>>>>>>>>> 发件人: Artem Ervits
>>>>>>>>>>>>>>>>>>>>> 发送时间: 2012-08-09  22:16:53
>>>>>>>>>>>>>>>>>>>>> 收件人: [email protected]
>>>>>>>>>>>>>>>>>>>>> 抄送:
>>>>>>>>>>>>>>>>>>>>> 主题: RE: Problem when setting up hadoop cluster step 2
>>>>>>>>>>>>>>>>>>>>> the installer requires a hosts file which I believe you 
>>>>>>>>>>>>>>>>>>>>> called hostdetail. Make sure it's the same file. You also 
>>>>>>>>>>>>>>>>>>>>> mention a hosts.txt and host.txt. You only need one file 
>>>>>>>>>>>>>>>>>>>>> with the names of all servers.
>>>>>>>>>>>>>>>>>>>>> -----Original Message-----
>>>>>>>>>>>>>>>>>>>>> From: xu peng [mailto:[email protected]]
>>>>>>>>>>>>>>>>>>>>> Sent: Thursday, August 09, 2012 2:02 AM
>>>>>>>>>>>>>>>>>>>>> To: [email protected]
>>>>>>>>>>>>>>>>>>>>> Subject: Problem when setting up hadoop cluster step 2
>>>>>>>>>>>>>>>>>>>>> Hi everyone :
>>>>>>>>>>>>>>>>>>>>> I am trying to use ambari to set up a hadoop cluster , 
>>>>>>>>>>>>>>>>>>>>> but i encounter a problem on step 2. I already set up the 
>>>>>>>>>>>>>>>>>>>>> password-less ssh, and i creat a hostdetail.txt file.
>>>>>>>>>>>>>>>>>>>>> The problem is that i found the file
>>>>>>>>>>>>>>>>>>>>> "/var/run/hmc/clusters/EBHadoop/hosts.txt" is empty , no 
>>>>>>>>>>>>>>>>>>>>> matter how many times i submit the host.txt file on the 
>>>>>>>>>>>>>>>>>>>>> website , and i really don't know why.
>>>>>>>>>>>>>>>>>>>>> {
>>>>>>>>>>>>>>>>>>>>> Here is the log file : [2012:08:09
>>>>>>>>>>>>>>>>>>>>> 05:17:56][ERROR][sequentialScriptExecutor][sequentialScriptRunner.php:272][]:
>>>>>>>>>>>>>>>>>>>>> Encountered total failure in transaction 100 while 
>>>>>>>>>>>>>>>>>>>>> running cmd:
>>>>>>>>>>>>>>>>>>>>> /usr/bin/php ./addNodes/findSshableNodes.php with args: 
>>>>>>>>>>>>>>>>>>>>> EBHadoop root
>>>>>>>>>>>>>>>>>>>>> 35 100 36 /var/run/hmc/clusters/EBHadoop/hosts.txt
>>>>>>>>>>>>>>>>>>>>> and my host.txt is like this(vbaby1.cloud.eb is the 
>>>>>>>>>>>>>>>>>>>>> master node) :
>>>>>>>>>>>>>>>>>>>>> vbaby2.cloud.eb
>>>>>>>>>>>>>>>>>>>>> vbaby3.cloud.eb
>>>>>>>>>>>>>>>>>>>>> vbaby4.cloud.eb
>>>>>>>>>>>>>>>>>>>>> vbaby5.cloud.eb
>>>>>>>>>>>>>>>>>>>>> vbaby6.cloud.eb
>>>>>>>>>>>>>>>>>>>>> }
>>>>>>>>>>>>>>>>>>>>> Can anyone help me and tell me what i am doing wrong ?
>>>>>>>>>>>>>>>>>>>>> Thank you very much ~!
>>>>>>>>>>>>>>>>>>>>> This electronic message is intended to be for the use 
>>>>>>>>>>>>>>>>>>>>> only of the named recipient, and may contain information 
>>>>>>>>>>>>>>>>>>>>> that is confidential or privileged. If you are not the 
>>>>>>>>>>>>>>>>>>>>> intended recipient, you are hereby notified that any 
>>>>>>>>>>>>>>>>>>>>> disclosure, copying, distribution or use of the contents 
>>>>>>>>>>>>>>>>>>>>> of this message is strictly prohibited. If you have 
>>>>>>>>>>>>>>>>>>>>> received this message in error or are not the named 
>>>>>>>>>>>>>>>>>>>>> recipient, please notify us immediately by contacting the 
>>>>>>>>>>>>>>>>>>>>> sender at the electronic mail address noted above, and 
>>>>>>>>>>>>>>>>>>>>> delete and destroy all copies of this message. Thank you.
>>>>>>>>>>>>>>>>>>>>> --------------------
>>>>>>>>>>>>>>>>>>>>> This electronic message is intended to be for the use 
>>>>>>>>>>>>>>>>>>>>> only of the named recipient, and may contain information 
>>>>>>>>>>>>>>>>>>>>> that is confidential or privileged.  If you are not the 
>>>>>>>>>>>>>>>>>>>>> intended recipient, you are hereby notified that any 
>>>>>>>>>>>>>>>>>>>>> disclosure, copying, distribution or use of the contents 
>>>>>>>>>>>>>>>>>>>>> of this message is strictly prohibited.  If you have 
>>>>>>>>>>>>>>>>>>>>> received this message in error or are not the named 
>>>>>>>>>>>>>>>>>>>>> recipient, please notify us immediately by contacting the 
>>>>>>>>>>>>>>>>>>>>> sender at the electronic mail address noted above, and 
>>>>>>>>>>>>>>>>>>>>> delete and destroy all copies of this message.  Thank you.
>>>>>>>>>>>>>>>>>>>>> --------------------
>>>>>>>>>>>>>>>>>>>>> This electronic message is intended to be for the use 
>>>>>>>>>>>>>>>>>>>>> only of the named recipient, and may contain information 
>>>>>>>>>>>>>>>>>>>>> that is confidential or privileged.  If you are not the 
>>>>>>>>>>>>>>>>>>>>> intended recipient, you are hereby notified that any 
>>>>>>>>>>>>>>>>>>>>> disclosure, copying, distribution or use of the contents 
>>>>>>>>>>>>>>>>>>>>> of this message is strictly prohibited.  If you have 
>>>>>>>>>>>>>>>>>>>>> received this message in error or are not the named 
>>>>>>>>>>>>>>>>>>>>> recipient, please notify us immediately by contacting the 
>>>>>>>>>>>>>>>>>>>>> sender at the electronic mail address noted above, and 
>>>>>>>>>>>>>>>>>>>>> delete and destroy all copies of this message.  Thank you.
>>>>>>>>>>>>>>>>>>>>> --------------------
>>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>>> This electronic message is intended to be for the use 
>>>>>>>>>>>>>>>>>>>>> only of the named recipient, and may contain information 
>>>>>>>>>>>>>>>>>>>>> that is confidential or privileged.  If you are not the 
>>>>>>>>>>>>>>>>>>>>> intended recipient, you are hereby notified that any 
>>>>>>>>>>>>>>>>>>>>> disclosure, copying, distribution or use of the contents 
>>>>>>>>>>>>>>>>>>>>> of this message is strictly prohibited.  If you have 
>>>>>>>>>>>>>>>>>>>>> received this message in error or are not the named 
>>>>>>>>>>>>>>>>>>>>> recipient, please notify us immediately by contacting the 
>>>>>>>>>>>>>>>>>>>>> sender at the electronic mail address noted above, and 
>>>>>>>>>>>>>>>>>>>>> delete and destroy all copies of this message.  Thank you.
>>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>>> --------------------
>>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>>> This electronic message is intended to be for the use 
>>>>>>>>>>>>>>>>>>>>> only of the named recipient, and may contain information 
>>>>>>>>>>>>>>>>>>>>> that is confidential or privileged.  If you are not the 
>>>>>>>>>>>>>>>>>>>>> intended recipient, you are hereby notified that any 
>>>>>>>>>>>>>>>>>>>>> disclosure, copying, distribution or use of the contents 
>>>>>>>>>>>>>>>>>>>>> of this message is strictly prohibited.  If you have 
>>>>>>>>>>>>>>>>>>>>> received this message in error or are not the named 
>>>>>>>>>>>>>>>>>>>>> recipient, please notify us immediately by contacting the 
>>>>>>>>>>>>>>>>>>>>> sender at the electronic mail address noted above, and 
>>>>>>>>>>>>>>>>>>>>> delete and destroy all copies of this message.  Thank you.
>>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> <hmcLog.txt><hostdetails.txt><httpdLog.txt><ssh1.sh><ssh1_result.jpg>
>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> 
>>>>>>>>>>>> 
>>>>>>>>>>> <DeployError1_2012.8.13.txt><log.rar><DeployingDetails.jpg><UninstallError1_2012.8.13.txt>
>>>>>>>>>> 
>>>>>>>> 
>>>>>>> <deployError2012.8.17.txt>
>>>>>> 
>>>> 
> <gangliaStartError.txt><4.jpg>

Reply via email to