Hi Hitesh :

I use the default settings of the mount point , but it seems like this
path is not a directory(/dev/mapper/hdvg-rootlv/), and i can not
execute mkdir -p command on this path. And the hdvg-rootlv is a
blocking file (bwrxwrxwrx) .  Is there something wrong ?



On Sun, Aug 19, 2012 at 3:38 AM, Hitesh Shah <[email protected]> wrote:
> Hi
>
> Yes - you should all packages from the new repo and none from the old repo. 
> Most of the packages should be the same but same like hadoop-lzo were 
> re-factored to work correctly with respect to 32/64-bit installs on RHEL6.
>
> Regarding the mount points, from a hadoop point of view, the namenode and 
> datanode dirs are just dirs. From a performance point of view, you want each 
> dir to be created on a separate mount point to increase disk io bandwidth. 
> This means that the mount points that you select on the UI should allow 
> directories to be created. If you have mounted certain kind of filesystems 
> which you do not wish to use for hadoop ( any tmpfs, nfs mounts etc ), you 
> should de-select them on the UI and/or use the custom mount point text box as 
> appropriate. The UI currently does not distinguish valid mount points and 
> therefore it is up to the user to select correctly.
>
> -- Hitesh
>
>
> On Aug 18, 2012, at 9:48 AM, xu peng wrote:
>
>> Hi Hitesh:
>>
>> Thanks again for your reply.
>>
>> I solved the dependency problem after updating the hdp repo.
>>
>> But here comes two new problems :
>> 1. I update the new hdp repo , but i create a local repo copy of the
>> old hdp repo. And I installed all the rpm package except
>> hadoop-lzo-native using the old hdp repo. So it seems like the
>> hadoop-lzo-native has some conflct with hadoop-lzo. So , do i have to
>> install all the rpm package from the new repo ?
>>
>> 2. From the error log , i can see a command "mkdir -p /var/.../..
>> (mounting point of hadoop)", but i found the mouting point is not a
>> dir , but a blocking file(bwrxwrxwrx). And the execution of this step
>> failed. Did i do something wrong ?
>>
>> I am sorry that this deploy error log is on my company's computer, and
>> i will upload it in my next email.
>>
>>
>> Thanks
>> -- Xupeng
>>
>> On Sat, Aug 18, 2012 at 4:43 AM, Hitesh Shah <[email protected]> wrote:
>>> Hi again,
>>>
>>> You are actually hitting a problem caused by some changes in the code which 
>>> require a modified repo. Unfortunately, I got delayed in modifying the 
>>> documentation to point to the new repo.
>>>
>>> Could you try using
>>> http://public-repo-1.hortonworks.com/HDP-1.0.1.14/repos/centos5/hdp-release-1.0.1.14-1.el5.noarch.rpm
>>> or
>>> http://public-repo-1.hortonworks.com/HDP-1.0.1.14/repos/centos6/hdp-release-1.0.1.14-1.el6.noarch.rpm
>>>
>>> The above should install the yum repo configs to point to the correct repo 
>>> which will have the lzo packages.
>>>
>>> -- Hitesh
>>>
>>>
>>> On Aug 16, 2012, at 9:27 PM, xu peng wrote:
>>>
>>>> Hitesh Shah :
>>>>
>>>> It is a my my pleasure to fill jira of ambari to help other users . As
>>>> a matter of fact, i want to summarize all the problem before i install
>>>> ambari cluster successfully. And i will  feed back as soon as
>>>> possiable.
>>>>
>>>> Here is another problem i encounter when install hadoop using ambari,
>>>> i found a rpm package "hadoop-lzp-native" not in the hdp repo
>>>> (baseurl=http://public-repo-1.hortonworks.com/HDP-1.0.13/repos/centos5)
>>>> . So i failed againg during deploying step.
>>>>
>>>> And the attachment is the deploying log , please refer.
>>>>
>>>> Thanks a lot and look forward to you reply.
>>>>
>>>>
>>>> On Tue, Aug 14, 2012 at 11:35 PM, Hitesh Shah <[email protected]> 
>>>> wrote:
>>>>> Ok - the cert issue is sometimes a result of uninstalling and 
>>>>> re-installing ambari agents.
>>>>>
>>>>> The re-install causes ambari agents to regenerate a new certification and 
>>>>> if the master was bootstrapped earlier, it would still be looking to 
>>>>> match against old certs.
>>>>>
>>>>> Stop ambari master and remove ambari-agent rpm from all hosts.
>>>>>
>>>>> To fix this:
>>>>>  - on the master, do a puppet cert revoke for all hosts ( 
>>>>> http://docs.puppetlabs.com/man/cert.html )
>>>>>  - you can do a cert list to get all signed or non-signed hosts
>>>>>
>>>>> On all hosts, delete the following dirs ( if they exist ) :
>>>>>  - /etc/puppet/ssl
>>>>>  - /etc/puppet/[master|agent\/ssl/
>>>>>  - /var/lib/puppet/ssl/
>>>>>
>>>>>
>>>>> After doing the above, re-install the ambari agent.
>>>>>
>>>>> On the ambari master, stop the master. Run the following command:
>>>>>
>>>>> puppet master --no-daemonize --debug
>>>>>
>>>>> The above runs in the foreground. The reason to run this is to make sure 
>>>>> the cert for the master is recreated as we deleted it earlier.
>>>>>
>>>>> Now, kill the above process running in the foreground and do a service 
>>>>> ambari start to bring up the UI.
>>>>>
>>>>> You should be able to bootstrap from this point on.
>>>>>
>>>>> Would you mind filing a jira and mentioning all the various issues you 
>>>>> have come across and how you solved them. We can use that to create an 
>>>>> FAQ for other users.
>>>>>
>>>>> thanks
>>>>> -- Hitesh
>>>>>
>>>>>
>>>>> On Aug 14, 2012, at 1:55 AM, xu peng wrote:
>>>>>
>>>>>> Hi  Hitesh :
>>>>>>
>>>>>> Thanks a lot for your reply.
>>>>>>
>>>>>> 1. I did a puppet kick --ping to the client from my ambari master ,
>>>>>> all the five nodes failed with the same log (Triggering
>>>>>> vbaby2.cloud.eb
>>>>>> Host vbaby2.cloud.eb failed: certificate verify failed.  This is often
>>>>>> because the time is out of sync on the server or client
>>>>>> vbaby2.cloud.eb finished with exit code 2)
>>>>>>
>>>>>> I manually run "service ambari-agent start" , is that necessary ? How
>>>>>> can i fix these problem ?
>>>>>>
>>>>>> 2. As you suggest , I run the yum command manually. And found that the
>>>>>> installation missed some dependecy - php-gd. And i have to update my
>>>>>> yum repo.
>>>>>>
>>>>>>
>>>>>>
>>>>>> On Tue, Aug 14, 2012 at 1:01 AM, Hitesh Shah <[email protected]> 
>>>>>> wrote:
>>>>>>> Based on your deploy error log:
>>>>>>>
>>>>>>> "3": {
>>>>>>>      "nodeReport": {
>>>>>>>          "PUPPET_KICK_FAILED": [],
>>>>>>>          "PUPPET_OPERATION_FAILED": [
>>>>>>>              "vbaby3.cloud.eb",
>>>>>>>              "vbaby5.cloud.eb",
>>>>>>>              "vbaby4.cloud.eb",
>>>>>>>              "vbaby2.cloud.eb",
>>>>>>>              "vbaby6.cloud.eb",
>>>>>>>              "vbaby1.cloud.eb"
>>>>>>>          ],
>>>>>>>          "PUPPET_OPERATION_TIMEDOUT": [
>>>>>>>              "vbaby5.cloud.eb",
>>>>>>>              "vbaby4.cloud.eb",
>>>>>>>              "vbaby2.cloud.eb",
>>>>>>>              "vbaby6.cloud.eb",
>>>>>>>              "vbaby1.cloud.eb"
>>>>>>>          ],
>>>>>>>
>>>>>>> 5 nodes timed out which means the puppet agent is not running on them 
>>>>>>> or they cannot communicate with the master. Trying doing a puppet kick 
>>>>>>> --ping to them from the master.
>>>>>>>
>>>>>>> For the one which failed, it failed at
>>>>>>>
>>>>>>> "\"Mon Aug 13 11:54:17 +0800 2012 
>>>>>>> /Stage[1]/Hdp::Pre_install_pkgs/Hdp::Exec[yum install 
>>>>>>> $pre_installed_pkgs]/Exec[yum install $pre_installed_pkgs]/returns 
>>>>>>> (err): change from notrun to 0 failed: yum install -y hadoop 
>>>>>>> hadoop-libhdfs hadoop-native hadoop-pipes hadoop-sbin hadoop-lzo hadoop 
>>>>>>> hadoop-libhdfs hadoop-native hadoop-pipes hadoop-sbin hadoop-lzo 
>>>>>>> hdp_mon_dashboard ganglia-gmond-3.2.0 gweb hdp_mon_ganglia_addons 
>>>>>>> snappy snappy-devel returned 1 instead of one of [0] at 
>>>>>>> /etc/puppet/agent/modules/hdp/manifests/init.pp:265\"",
>>>>>>>
>>>>>>> It seems like yum install failed on the host. Try running the command 
>>>>>>> manually and see what the error is.
>>>>>>>
>>>>>>> -- Hitesh
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> On Aug 13, 2012, at 2:28 AM, xu peng wrote:
>>>>>>>
>>>>>>>> Hi Hitesh :
>>>>>>>>
>>>>>>>> It's me again.
>>>>>>>>
>>>>>>>> Followed you advice , I reinstalled the ambari server. But deploying
>>>>>>>> cluster and uninstall cluster failed again. I really  don't know why.
>>>>>>>>
>>>>>>>> I supplied a attachment which contains the logs of  all the nodes in
>>>>>>>> my cluster (/var/log/puppet_*.log , /var/log/puppet/*.log ,
>>>>>>>> /var/log/yum.log, /var/log/hmc/hmc.log). And vbaby3.cloud.eb is the
>>>>>>>> ambari server. Please refer.
>>>>>>>>
>>>>>>>> Attachment DeployError and UninstallError is the log supplied by the
>>>>>>>> website of ambari when failing. And attachment DeployingDetails.jpg is
>>>>>>>> the deploy details of my cluster. Please refer.
>>>>>>>>
>>>>>>>>
>>>>>>>> Thanks again for your patience ! And look forward to your reply.
>>>>>>>>
>>>>>>>> Xupeng
>>>>>>>>
>>>>>>>> On Sat, Aug 11, 2012 at 10:56 PM, Hitesh Shah <[email protected]> 
>>>>>>>> wrote:
>>>>>>>>> For uninstall failures, you will need to do a couple of things. 
>>>>>>>>> Depending on where the uninstall failed, you may have to manually do 
>>>>>>>>> a killall java on all the nodes to kill any missed processes. If you 
>>>>>>>>> want to start with a complete clean install, you should also delete 
>>>>>>>>> the hadoop dir in the mount points you selected during the previous 
>>>>>>>>> install  so that the new fresh install does not face errors when it 
>>>>>>>>> tries to re-format hdfs.
>>>>>>>>>
>>>>>>>>> After that, simply, uinstall and re-install ambari rpm and that 
>>>>>>>>> should allow you to re-create a fresh cluster.
>>>>>>>>>
>>>>>>>>> -- Hitesh
>>>>>>>>>
>>>>>>>>> On Aug 11, 2012, at 2:34 AM, xu peng wrote:
>>>>>>>>>
>>>>>>>>>> Hi Hitesh :
>>>>>>>>>>
>>>>>>>>>> Thanks a lot for your reply.
>>>>>>>>>>
>>>>>>>>>> I solved this problem , it is silly mistake. Someone has changed the
>>>>>>>>>> owner of "/" dir , and according to the errorlog , pdsh need root to
>>>>>>>>>> proceed.
>>>>>>>>>>
>>>>>>>>>> After changing the owner of "/" to root , problem solved. Thank you
>>>>>>>>>> again for you reply.
>>>>>>>>>>
>>>>>>>>>> I have another question. I had a uninstall failure , and there is no
>>>>>>>>>> button on the website for me to rollback and i don't know what to do
>>>>>>>>>> about that. What should i do now to reinstall hadoop ?
>>>>>>>>>>
>>>>>>>>>> Thanks
>>>>>>>>>>
>>>>>>>>>> On Fri, Aug 10, 2012 at 10:55 PM, Hitesh Shah 
>>>>>>>>>> <[email protected]> wrote:
>>>>>>>>>>> Hi
>>>>>>>>>>>
>>>>>>>>>>> Currently, the ambari installer requires everything to be run as 
>>>>>>>>>>> root. It does not detect that the user is not root and use sudo 
>>>>>>>>>>> either on the master or on the agent nodes.
>>>>>>>>>>> Furthermore, it seems like it is failing when trying to use pdsh to 
>>>>>>>>>>> make remote calls to the host list that you passed in due to the 
>>>>>>>>>>> errors mentioned in your script. This could be due to how it was 
>>>>>>>>>>> installed but I am not sure.
>>>>>>>>>>>
>>>>>>>>>>> Could you switch to become root and run any simple command on all 
>>>>>>>>>>> hosts using pdsh? If you want to reference exactly how ambari uses 
>>>>>>>>>>> pdsh, you can look into /usr/share/hmc/php/frontend/commandUtils.php
>>>>>>>>>>>
>>>>>>>>>>> thanks
>>>>>>>>>>> -- Hitesh
>>>>>>>>>>>
>>>>>>>>>>> On Aug 9, 2012, at 9:04 PM, xu peng wrote:
>>>>>>>>>>>
>>>>>>>>>>>> According to the error log , is there something wrong with my 
>>>>>>>>>>>> account ?
>>>>>>>>>>>>
>>>>>>>>>>>> I installed all the dependency module and ambari with the user
>>>>>>>>>>>> "ambari" instead of root. I added user "ambari" to /etc/sudofilers
>>>>>>>>>>>> with no passwd.
>>>>>>>>>>>>
>>>>>>>>>>>> On Fri, Aug 10, 2012 at 11:49 AM, xu peng <[email protected]> 
>>>>>>>>>>>> wrote:
>>>>>>>>>>>>> There is no 100.log.file in /var/log/hmc dir, but only 55.log 
>>>>>>>>>>>>> file (55
>>>>>>>>>>>>> is the biggest version num).
>>>>>>>>>>>>>
>>>>>>>>>>>>> The content of 55.log is :
>>>>>>>>>>>>> pdsh@vbaby1: module path "/usr/lib64/pdsh" insecure.
>>>>>>>>>>>>> pdsh@vbaby1: "/": Owner not root, current uid, or pdsh executable 
>>>>>>>>>>>>> owner
>>>>>>>>>>>>> pdsh@vbaby1: Couldn't load any pdsh modules
>>>>>>>>>>>>>
>>>>>>>>>>>>> Thanks ~
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> On Fri, Aug 10, 2012 at 11:36 AM, Hitesh Shah 
>>>>>>>>>>>>> <[email protected]> wrote:
>>>>>>>>>>>>>> Sorry - my mistake. The last txn mentioned is 100 so please look 
>>>>>>>>>>>>>> for the 100.log file.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> -- Hitesh
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> On Aug 9, 2012, at 8:34 PM, Hitesh Shah wrote:
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Thanks - will take a look and get back to you.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Could you also look at /var/log/hmc/hmc.txn.55.log and see if 
>>>>>>>>>>>>>>> there are any errors in it?
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> -- Hitesh.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> On Aug 9, 2012, at 8:00 PM, xu peng wrote:
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Hi Hitesh :
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Thanks a lot for your replying. I have done all your 
>>>>>>>>>>>>>>>> suggestions in my
>>>>>>>>>>>>>>>> ambari server , and the result is as below.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> 1. I can confirm that the hosts.txt file is empty after i 
>>>>>>>>>>>>>>>> failed at
>>>>>>>>>>>>>>>> the step finding reachable nodes.
>>>>>>>>>>>>>>>> 2. I tried make hostdetails file in win7 and redhat , it both
>>>>>>>>>>>>>>>> failed.(Please see the attachment, my hostdetails file)
>>>>>>>>>>>>>>>> 3. I removed the logging re-direct and run the .sh script .It 
>>>>>>>>>>>>>>>> seems
>>>>>>>>>>>>>>>> like the script works well , it print the hostname in console 
>>>>>>>>>>>>>>>> and
>>>>>>>>>>>>>>>> generate a file (content  is "0") in the same dir. (Please see 
>>>>>>>>>>>>>>>> the
>>>>>>>>>>>>>>>> attachment , the result and my .sh script )
>>>>>>>>>>>>>>>> 4. I attached the hmc.log and error_log too. Hope this helps ~
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Thanks ~
>>>>>>>>>>>>>>>> Xupeng
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> On Fri, Aug 10, 2012 at 12:24 AM, Hitesh Shah 
>>>>>>>>>>>>>>>> <[email protected]> wrote:
>>>>>>>>>>>>>>>>> Xupeng, can you confirm that the hosts.txt file at 
>>>>>>>>>>>>>>>>> /var/run/hmc/clusters/EBHadoop/hosts.txt is empty?
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Also, can you ensure that the hostdetails file that you 
>>>>>>>>>>>>>>>>> upload does not have any special characters that may be 
>>>>>>>>>>>>>>>>> creating problems for the parsing layer?
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> In the same dir, there should be an ssh.sh script. Can you 
>>>>>>>>>>>>>>>>> create a copy of it, edit to remove the logging re-directs to 
>>>>>>>>>>>>>>>>> files and run the script manually from command-line ( it 
>>>>>>>>>>>>>>>>> takes in a hostname as the argument ) ? The output of that 
>>>>>>>>>>>>>>>>> should show you as to what is going wrong.
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Also, please look at /var/log/hmc/hmc.log and httpd/error_log 
>>>>>>>>>>>>>>>>> to see if there are any errors being logged which may shed 
>>>>>>>>>>>>>>>>> more light on the issue.
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> thanks
>>>>>>>>>>>>>>>>> -- Hitesh
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> On Aug 9, 2012, at 9:11 AM, Artem Ervits wrote:
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Which file are you supplying in the step? Hostdetail.txt or 
>>>>>>>>>>>>>>>>>> hosts?
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> From: xupeng.bupt [mailto:[email protected]]
>>>>>>>>>>>>>>>>>> Sent: Thursday, August 09, 2012 11:33 AM
>>>>>>>>>>>>>>>>>> To: ambari-user
>>>>>>>>>>>>>>>>>> Subject: Re: RE: Problem when setting up hadoop cluster step 
>>>>>>>>>>>>>>>>>> 2
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Thank you for your replying ~
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> I made only one hostdetail.txt file which contains the names 
>>>>>>>>>>>>>>>>>> of all servers. And i submit this file on the website ,  but 
>>>>>>>>>>>>>>>>>> i still have the same problem. I failed at the step of 
>>>>>>>>>>>>>>>>>> finding reachable nodes.
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> The error log is : "
>>>>>>>>>>>>>>>>>> [ERROR][sequentialScriptExecutor][sequentialScriptRunner.php:272][]:
>>>>>>>>>>>>>>>>>> Encountered total failure in transaction 100 while running 
>>>>>>>>>>>>>>>>>> cmd:
>>>>>>>>>>>>>>>>>> /usr/bin/php ./addNodes/findSshableNodes.php with args: 
>>>>>>>>>>>>>>>>>> EBHadoop root
>>>>>>>>>>>>>>>>>> 35 100 36 /var/run/hmc/clusters/EBHadoop/hosts.txt
>>>>>>>>>>>>>>>>>> "
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> And my hostdetail.txt file is :"
>>>>>>>>>>>>>>>>>> vbaby2.cloud.eb
>>>>>>>>>>>>>>>>>> vbaby3.cloud.eb
>>>>>>>>>>>>>>>>>> vbaby4.cloud.eb
>>>>>>>>>>>>>>>>>> vbaby5.cloud.eb
>>>>>>>>>>>>>>>>>> vbaby6.cloud.eb
>>>>>>>>>>>>>>>>>> "
>>>>>>>>>>>>>>>>>> Thank you very much ~
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> 2012-08-09
>>>>>>>>>>>>>>>>>> xupeng.bupt
>>>>>>>>>>>>>>>>>> 发件人: Artem Ervits
>>>>>>>>>>>>>>>>>> 发送时间: 2012-08-09  22:16:53
>>>>>>>>>>>>>>>>>> 收件人: [email protected]
>>>>>>>>>>>>>>>>>> 抄送:
>>>>>>>>>>>>>>>>>> 主题: RE: Problem when setting up hadoop cluster step 2
>>>>>>>>>>>>>>>>>> the installer requires a hosts file which I believe you 
>>>>>>>>>>>>>>>>>> called hostdetail. Make sure it's the same file. You also 
>>>>>>>>>>>>>>>>>> mention a hosts.txt and host.txt. You only need one file 
>>>>>>>>>>>>>>>>>> with the names of all servers.
>>>>>>>>>>>>>>>>>> -----Original Message-----
>>>>>>>>>>>>>>>>>> From: xu peng [mailto:[email protected]]
>>>>>>>>>>>>>>>>>> Sent: Thursday, August 09, 2012 2:02 AM
>>>>>>>>>>>>>>>>>> To: [email protected]
>>>>>>>>>>>>>>>>>> Subject: Problem when setting up hadoop cluster step 2
>>>>>>>>>>>>>>>>>> Hi everyone :
>>>>>>>>>>>>>>>>>> I am trying to use ambari to set up a hadoop cluster , but i 
>>>>>>>>>>>>>>>>>> encounter a problem on step 2. I already set up the 
>>>>>>>>>>>>>>>>>> password-less ssh, and i creat a hostdetail.txt file.
>>>>>>>>>>>>>>>>>> The problem is that i found the file
>>>>>>>>>>>>>>>>>> "/var/run/hmc/clusters/EBHadoop/hosts.txt" is empty , no 
>>>>>>>>>>>>>>>>>> matter how many times i submit the host.txt file on the 
>>>>>>>>>>>>>>>>>> website , and i really don't know why.
>>>>>>>>>>>>>>>>>> {
>>>>>>>>>>>>>>>>>> Here is the log file : [2012:08:09
>>>>>>>>>>>>>>>>>> 05:17:56][ERROR][sequentialScriptExecutor][sequentialScriptRunner.php:272][]:
>>>>>>>>>>>>>>>>>> Encountered total failure in transaction 100 while running 
>>>>>>>>>>>>>>>>>> cmd:
>>>>>>>>>>>>>>>>>> /usr/bin/php ./addNodes/findSshableNodes.php with args: 
>>>>>>>>>>>>>>>>>> EBHadoop root
>>>>>>>>>>>>>>>>>> 35 100 36 /var/run/hmc/clusters/EBHadoop/hosts.txt
>>>>>>>>>>>>>>>>>> and my host.txt is like this(vbaby1.cloud.eb is the master 
>>>>>>>>>>>>>>>>>> node) :
>>>>>>>>>>>>>>>>>> vbaby2.cloud.eb
>>>>>>>>>>>>>>>>>> vbaby3.cloud.eb
>>>>>>>>>>>>>>>>>> vbaby4.cloud.eb
>>>>>>>>>>>>>>>>>> vbaby5.cloud.eb
>>>>>>>>>>>>>>>>>> vbaby6.cloud.eb
>>>>>>>>>>>>>>>>>> }
>>>>>>>>>>>>>>>>>> Can anyone help me and tell me what i am doing wrong ?
>>>>>>>>>>>>>>>>>> Thank you very much ~!
>>>>>>>>>>>>>>>>>> This electronic message is intended to be for the use only 
>>>>>>>>>>>>>>>>>> of the named recipient, and may contain information that is 
>>>>>>>>>>>>>>>>>> confidential or privileged. If you are not the intended 
>>>>>>>>>>>>>>>>>> recipient, you are hereby notified that any disclosure, 
>>>>>>>>>>>>>>>>>> copying, distribution or use of the contents of this message 
>>>>>>>>>>>>>>>>>> is strictly prohibited. If you have received this message in 
>>>>>>>>>>>>>>>>>> error or are not the named recipient, please notify us 
>>>>>>>>>>>>>>>>>> immediately by contacting the sender at the electronic mail 
>>>>>>>>>>>>>>>>>> address noted above, and delete and destroy all copies of 
>>>>>>>>>>>>>>>>>> this message. Thank you.
>>>>>>>>>>>>>>>>>> --------------------
>>>>>>>>>>>>>>>>>> This electronic message is intended to be for the use only 
>>>>>>>>>>>>>>>>>> of the named recipient, and may contain information that is 
>>>>>>>>>>>>>>>>>> confidential or privileged.  If you are not the intended 
>>>>>>>>>>>>>>>>>> recipient, you are hereby notified that any disclosure, 
>>>>>>>>>>>>>>>>>> copying, distribution or use of the contents of this message 
>>>>>>>>>>>>>>>>>> is strictly prohibited.  If you have received this message 
>>>>>>>>>>>>>>>>>> in error or are not the named recipient, please notify us 
>>>>>>>>>>>>>>>>>> immediately by contacting the sender at the electronic mail 
>>>>>>>>>>>>>>>>>> address noted above, and delete and destroy all copies of 
>>>>>>>>>>>>>>>>>> this message.  Thank you.
>>>>>>>>>>>>>>>>>> --------------------
>>>>>>>>>>>>>>>>>> This electronic message is intended to be for the use only 
>>>>>>>>>>>>>>>>>> of the named recipient, and may contain information that is 
>>>>>>>>>>>>>>>>>> confidential or privileged.  If you are not the intended 
>>>>>>>>>>>>>>>>>> recipient, you are hereby notified that any disclosure, 
>>>>>>>>>>>>>>>>>> copying, distribution or use of the contents of this message 
>>>>>>>>>>>>>>>>>> is strictly prohibited.  If you have received this message 
>>>>>>>>>>>>>>>>>> in error or are not the named recipient, please notify us 
>>>>>>>>>>>>>>>>>> immediately by contacting the sender at the electronic mail 
>>>>>>>>>>>>>>>>>> address noted above, and delete and destroy all copies of 
>>>>>>>>>>>>>>>>>> this message.  Thank you.
>>>>>>>>>>>>>>>>>> --------------------
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> This electronic message is intended to be for the use only 
>>>>>>>>>>>>>>>>>> of the named recipient, and may contain information that is 
>>>>>>>>>>>>>>>>>> confidential or privileged.  If you are not the intended 
>>>>>>>>>>>>>>>>>> recipient, you are hereby notified that any disclosure, 
>>>>>>>>>>>>>>>>>> copying, distribution or use of the contents of this message 
>>>>>>>>>>>>>>>>>> is strictly prohibited.  If you have received this message 
>>>>>>>>>>>>>>>>>> in error or are not the named recipient, please notify us 
>>>>>>>>>>>>>>>>>> immediately by contacting the sender at the electronic mail 
>>>>>>>>>>>>>>>>>> address noted above, and delete and destroy all copies of 
>>>>>>>>>>>>>>>>>> this message.  Thank you.
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> --------------------
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> This electronic message is intended to be for the use only 
>>>>>>>>>>>>>>>>>> of the named recipient, and may contain information that is 
>>>>>>>>>>>>>>>>>> confidential or privileged.  If you are not the intended 
>>>>>>>>>>>>>>>>>> recipient, you are hereby notified that any disclosure, 
>>>>>>>>>>>>>>>>>> copying, distribution or use of the contents of this message 
>>>>>>>>>>>>>>>>>> is strictly prohibited.  If you have received this message 
>>>>>>>>>>>>>>>>>> in error or are not the named recipient, please notify us 
>>>>>>>>>>>>>>>>>> immediately by contacting the sender at the electronic mail 
>>>>>>>>>>>>>>>>>> address noted above, and delete and destroy all copies of 
>>>>>>>>>>>>>>>>>> this message.  Thank you.
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> <hmcLog.txt><hostdetails.txt><httpdLog.txt><ssh1.sh><ssh1_result.jpg>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>
>>>>>>>> <DeployError1_2012.8.13.txt><log.rar><DeployingDetails.jpg><UninstallError1_2012.8.13.txt>
>>>>>>>
>>>>>
>>>> <deployError2012.8.17.txt>
>>>
>

Reply via email to