Hi Hitesh: Thanks again for your reply.
I solved the dependency problem after updating the hdp repo. But here comes two new problems : 1. I update the new hdp repo , but i create a local repo copy of the old hdp repo. And I installed all the rpm package except hadoop-lzo-native using the old hdp repo. So it seems like the hadoop-lzo-native has some conflct with hadoop-lzo. So , do i have to install all the rpm package from the new repo ? 2. From the error log , i can see a command "mkdir -p /var/.../.. (mounting point of hadoop)", but i found the mouting point is not a dir , but a blocking file(bwrxwrxwrx). And the execution of this step failed. Did i do something wrong ? I am sorry that this deploy error log is on my company's computer, and i will upload it in my next email. Thanks -- Xupeng On Sat, Aug 18, 2012 at 4:43 AM, Hitesh Shah <[email protected]> wrote: > Hi again, > > You are actually hitting a problem caused by some changes in the code which > require a modified repo. Unfortunately, I got delayed in modifying the > documentation to point to the new repo. > > Could you try using > http://public-repo-1.hortonworks.com/HDP-1.0.1.14/repos/centos5/hdp-release-1.0.1.14-1.el5.noarch.rpm > or > http://public-repo-1.hortonworks.com/HDP-1.0.1.14/repos/centos6/hdp-release-1.0.1.14-1.el6.noarch.rpm > > The above should install the yum repo configs to point to the correct repo > which will have the lzo packages. > > -- Hitesh > > > On Aug 16, 2012, at 9:27 PM, xu peng wrote: > >> Hitesh Shah : >> >> It is a my my pleasure to fill jira of ambari to help other users . As >> a matter of fact, i want to summarize all the problem before i install >> ambari cluster successfully. And i will feed back as soon as >> possiable. >> >> Here is another problem i encounter when install hadoop using ambari, >> i found a rpm package "hadoop-lzp-native" not in the hdp repo >> (baseurl=http://public-repo-1.hortonworks.com/HDP-1.0.13/repos/centos5) >> . So i failed againg during deploying step. >> >> And the attachment is the deploying log , please refer. >> >> Thanks a lot and look forward to you reply. >> >> >> On Tue, Aug 14, 2012 at 11:35 PM, Hitesh Shah <[email protected]> wrote: >>> Ok - the cert issue is sometimes a result of uninstalling and re-installing >>> ambari agents. >>> >>> The re-install causes ambari agents to regenerate a new certification and >>> if the master was bootstrapped earlier, it would still be looking to match >>> against old certs. >>> >>> Stop ambari master and remove ambari-agent rpm from all hosts. >>> >>> To fix this: >>> - on the master, do a puppet cert revoke for all hosts ( >>> http://docs.puppetlabs.com/man/cert.html ) >>> - you can do a cert list to get all signed or non-signed hosts >>> >>> On all hosts, delete the following dirs ( if they exist ) : >>> - /etc/puppet/ssl >>> - /etc/puppet/[master|agent\/ssl/ >>> - /var/lib/puppet/ssl/ >>> >>> >>> After doing the above, re-install the ambari agent. >>> >>> On the ambari master, stop the master. Run the following command: >>> >>> puppet master --no-daemonize --debug >>> >>> The above runs in the foreground. The reason to run this is to make sure >>> the cert for the master is recreated as we deleted it earlier. >>> >>> Now, kill the above process running in the foreground and do a service >>> ambari start to bring up the UI. >>> >>> You should be able to bootstrap from this point on. >>> >>> Would you mind filing a jira and mentioning all the various issues you have >>> come across and how you solved them. We can use that to create an FAQ for >>> other users. >>> >>> thanks >>> -- Hitesh >>> >>> >>> On Aug 14, 2012, at 1:55 AM, xu peng wrote: >>> >>>> Hi Hitesh : >>>> >>>> Thanks a lot for your reply. >>>> >>>> 1. I did a puppet kick --ping to the client from my ambari master , >>>> all the five nodes failed with the same log (Triggering >>>> vbaby2.cloud.eb >>>> Host vbaby2.cloud.eb failed: certificate verify failed. This is often >>>> because the time is out of sync on the server or client >>>> vbaby2.cloud.eb finished with exit code 2) >>>> >>>> I manually run "service ambari-agent start" , is that necessary ? How >>>> can i fix these problem ? >>>> >>>> 2. As you suggest , I run the yum command manually. And found that the >>>> installation missed some dependecy - php-gd. And i have to update my >>>> yum repo. >>>> >>>> >>>> >>>> On Tue, Aug 14, 2012 at 1:01 AM, Hitesh Shah <[email protected]> >>>> wrote: >>>>> Based on your deploy error log: >>>>> >>>>> "3": { >>>>> "nodeReport": { >>>>> "PUPPET_KICK_FAILED": [], >>>>> "PUPPET_OPERATION_FAILED": [ >>>>> "vbaby3.cloud.eb", >>>>> "vbaby5.cloud.eb", >>>>> "vbaby4.cloud.eb", >>>>> "vbaby2.cloud.eb", >>>>> "vbaby6.cloud.eb", >>>>> "vbaby1.cloud.eb" >>>>> ], >>>>> "PUPPET_OPERATION_TIMEDOUT": [ >>>>> "vbaby5.cloud.eb", >>>>> "vbaby4.cloud.eb", >>>>> "vbaby2.cloud.eb", >>>>> "vbaby6.cloud.eb", >>>>> "vbaby1.cloud.eb" >>>>> ], >>>>> >>>>> 5 nodes timed out which means the puppet agent is not running on them or >>>>> they cannot communicate with the master. Trying doing a puppet kick >>>>> --ping to them from the master. >>>>> >>>>> For the one which failed, it failed at >>>>> >>>>> "\"Mon Aug 13 11:54:17 +0800 2012 >>>>> /Stage[1]/Hdp::Pre_install_pkgs/Hdp::Exec[yum install >>>>> $pre_installed_pkgs]/Exec[yum install $pre_installed_pkgs]/returns (err): >>>>> change from notrun to 0 failed: yum install -y hadoop hadoop-libhdfs >>>>> hadoop-native hadoop-pipes hadoop-sbin hadoop-lzo hadoop hadoop-libhdfs >>>>> hadoop-native hadoop-pipes hadoop-sbin hadoop-lzo hdp_mon_dashboard >>>>> ganglia-gmond-3.2.0 gweb hdp_mon_ganglia_addons snappy snappy-devel >>>>> returned 1 instead of one of [0] at >>>>> /etc/puppet/agent/modules/hdp/manifests/init.pp:265\"", >>>>> >>>>> It seems like yum install failed on the host. Try running the command >>>>> manually and see what the error is. >>>>> >>>>> -- Hitesh >>>>> >>>>> >>>>> >>>>> On Aug 13, 2012, at 2:28 AM, xu peng wrote: >>>>> >>>>>> Hi Hitesh : >>>>>> >>>>>> It's me again. >>>>>> >>>>>> Followed you advice , I reinstalled the ambari server. But deploying >>>>>> cluster and uninstall cluster failed again. I really don't know why. >>>>>> >>>>>> I supplied a attachment which contains the logs of all the nodes in >>>>>> my cluster (/var/log/puppet_*.log , /var/log/puppet/*.log , >>>>>> /var/log/yum.log, /var/log/hmc/hmc.log). And vbaby3.cloud.eb is the >>>>>> ambari server. Please refer. >>>>>> >>>>>> Attachment DeployError and UninstallError is the log supplied by the >>>>>> website of ambari when failing. And attachment DeployingDetails.jpg is >>>>>> the deploy details of my cluster. Please refer. >>>>>> >>>>>> >>>>>> Thanks again for your patience ! And look forward to your reply. >>>>>> >>>>>> Xupeng >>>>>> >>>>>> On Sat, Aug 11, 2012 at 10:56 PM, Hitesh Shah <[email protected]> >>>>>> wrote: >>>>>>> For uninstall failures, you will need to do a couple of things. >>>>>>> Depending on where the uninstall failed, you may have to manually do a >>>>>>> killall java on all the nodes to kill any missed processes. If you want >>>>>>> to start with a complete clean install, you should also delete the >>>>>>> hadoop dir in the mount points you selected during the previous install >>>>>>> so that the new fresh install does not face errors when it tries to >>>>>>> re-format hdfs. >>>>>>> >>>>>>> After that, simply, uinstall and re-install ambari rpm and that should >>>>>>> allow you to re-create a fresh cluster. >>>>>>> >>>>>>> -- Hitesh >>>>>>> >>>>>>> On Aug 11, 2012, at 2:34 AM, xu peng wrote: >>>>>>> >>>>>>>> Hi Hitesh : >>>>>>>> >>>>>>>> Thanks a lot for your reply. >>>>>>>> >>>>>>>> I solved this problem , it is silly mistake. Someone has changed the >>>>>>>> owner of "/" dir , and according to the errorlog , pdsh need root to >>>>>>>> proceed. >>>>>>>> >>>>>>>> After changing the owner of "/" to root , problem solved. Thank you >>>>>>>> again for you reply. >>>>>>>> >>>>>>>> I have another question. I had a uninstall failure , and there is no >>>>>>>> button on the website for me to rollback and i don't know what to do >>>>>>>> about that. What should i do now to reinstall hadoop ? >>>>>>>> >>>>>>>> Thanks >>>>>>>> >>>>>>>> On Fri, Aug 10, 2012 at 10:55 PM, Hitesh Shah <[email protected]> >>>>>>>> wrote: >>>>>>>>> Hi >>>>>>>>> >>>>>>>>> Currently, the ambari installer requires everything to be run as >>>>>>>>> root. It does not detect that the user is not root and use sudo >>>>>>>>> either on the master or on the agent nodes. >>>>>>>>> Furthermore, it seems like it is failing when trying to use pdsh to >>>>>>>>> make remote calls to the host list that you passed in due to the >>>>>>>>> errors mentioned in your script. This could be due to how it was >>>>>>>>> installed but I am not sure. >>>>>>>>> >>>>>>>>> Could you switch to become root and run any simple command on all >>>>>>>>> hosts using pdsh? If you want to reference exactly how ambari uses >>>>>>>>> pdsh, you can look into /usr/share/hmc/php/frontend/commandUtils.php >>>>>>>>> >>>>>>>>> thanks >>>>>>>>> -- Hitesh >>>>>>>>> >>>>>>>>> On Aug 9, 2012, at 9:04 PM, xu peng wrote: >>>>>>>>> >>>>>>>>>> According to the error log , is there something wrong with my >>>>>>>>>> account ? >>>>>>>>>> >>>>>>>>>> I installed all the dependency module and ambari with the user >>>>>>>>>> "ambari" instead of root. I added user "ambari" to /etc/sudofilers >>>>>>>>>> with no passwd. >>>>>>>>>> >>>>>>>>>> On Fri, Aug 10, 2012 at 11:49 AM, xu peng <[email protected]> >>>>>>>>>> wrote: >>>>>>>>>>> There is no 100.log.file in /var/log/hmc dir, but only 55.log file >>>>>>>>>>> (55 >>>>>>>>>>> is the biggest version num). >>>>>>>>>>> >>>>>>>>>>> The content of 55.log is : >>>>>>>>>>> pdsh@vbaby1: module path "/usr/lib64/pdsh" insecure. >>>>>>>>>>> pdsh@vbaby1: "/": Owner not root, current uid, or pdsh executable >>>>>>>>>>> owner >>>>>>>>>>> pdsh@vbaby1: Couldn't load any pdsh modules >>>>>>>>>>> >>>>>>>>>>> Thanks ~ >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> On Fri, Aug 10, 2012 at 11:36 AM, Hitesh Shah >>>>>>>>>>> <[email protected]> wrote: >>>>>>>>>>>> Sorry - my mistake. The last txn mentioned is 100 so please look >>>>>>>>>>>> for the 100.log file. >>>>>>>>>>>> >>>>>>>>>>>> -- Hitesh >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> On Aug 9, 2012, at 8:34 PM, Hitesh Shah wrote: >>>>>>>>>>>> >>>>>>>>>>>>> Thanks - will take a look and get back to you. >>>>>>>>>>>>> >>>>>>>>>>>>> Could you also look at /var/log/hmc/hmc.txn.55.log and see if >>>>>>>>>>>>> there are any errors in it? >>>>>>>>>>>>> >>>>>>>>>>>>> -- Hitesh. >>>>>>>>>>>>> >>>>>>>>>>>>> On Aug 9, 2012, at 8:00 PM, xu peng wrote: >>>>>>>>>>>>> >>>>>>>>>>>>>> Hi Hitesh : >>>>>>>>>>>>>> >>>>>>>>>>>>>> Thanks a lot for your replying. I have done all your suggestions >>>>>>>>>>>>>> in my >>>>>>>>>>>>>> ambari server , and the result is as below. >>>>>>>>>>>>>> >>>>>>>>>>>>>> 1. I can confirm that the hosts.txt file is empty after i failed >>>>>>>>>>>>>> at >>>>>>>>>>>>>> the step finding reachable nodes. >>>>>>>>>>>>>> 2. I tried make hostdetails file in win7 and redhat , it both >>>>>>>>>>>>>> failed.(Please see the attachment, my hostdetails file) >>>>>>>>>>>>>> 3. I removed the logging re-direct and run the .sh script .It >>>>>>>>>>>>>> seems >>>>>>>>>>>>>> like the script works well , it print the hostname in console and >>>>>>>>>>>>>> generate a file (content is "0") in the same dir. (Please see >>>>>>>>>>>>>> the >>>>>>>>>>>>>> attachment , the result and my .sh script ) >>>>>>>>>>>>>> 4. I attached the hmc.log and error_log too. Hope this helps ~ >>>>>>>>>>>>>> >>>>>>>>>>>>>> Thanks ~ >>>>>>>>>>>>>> Xupeng >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> On Fri, Aug 10, 2012 at 12:24 AM, Hitesh Shah >>>>>>>>>>>>>> <[email protected]> wrote: >>>>>>>>>>>>>>> Xupeng, can you confirm that the hosts.txt file at >>>>>>>>>>>>>>> /var/run/hmc/clusters/EBHadoop/hosts.txt is empty? >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Also, can you ensure that the hostdetails file that you upload >>>>>>>>>>>>>>> does not have any special characters that may be creating >>>>>>>>>>>>>>> problems for the parsing layer? >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> In the same dir, there should be an ssh.sh script. Can you >>>>>>>>>>>>>>> create a copy of it, edit to remove the logging re-directs to >>>>>>>>>>>>>>> files and run the script manually from command-line ( it takes >>>>>>>>>>>>>>> in a hostname as the argument ) ? The output of that should >>>>>>>>>>>>>>> show you as to what is going wrong. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Also, please look at /var/log/hmc/hmc.log and httpd/error_log >>>>>>>>>>>>>>> to see if there are any errors being logged which may shed more >>>>>>>>>>>>>>> light on the issue. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> thanks >>>>>>>>>>>>>>> -- Hitesh >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> On Aug 9, 2012, at 9:11 AM, Artem Ervits wrote: >>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Which file are you supplying in the step? Hostdetail.txt or >>>>>>>>>>>>>>>> hosts? >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> From: xupeng.bupt [mailto:[email protected]] >>>>>>>>>>>>>>>> Sent: Thursday, August 09, 2012 11:33 AM >>>>>>>>>>>>>>>> To: ambari-user >>>>>>>>>>>>>>>> Subject: Re: RE: Problem when setting up hadoop cluster step 2 >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Thank you for your replying ~ >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> I made only one hostdetail.txt file which contains the names >>>>>>>>>>>>>>>> of all servers. And i submit this file on the website , but i >>>>>>>>>>>>>>>> still have the same problem. I failed at the step of finding >>>>>>>>>>>>>>>> reachable nodes. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> The error log is : " >>>>>>>>>>>>>>>> [ERROR][sequentialScriptExecutor][sequentialScriptRunner.php:272][]: >>>>>>>>>>>>>>>> Encountered total failure in transaction 100 while running cmd: >>>>>>>>>>>>>>>> /usr/bin/php ./addNodes/findSshableNodes.php with args: >>>>>>>>>>>>>>>> EBHadoop root >>>>>>>>>>>>>>>> 35 100 36 /var/run/hmc/clusters/EBHadoop/hosts.txt >>>>>>>>>>>>>>>> " >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> And my hostdetail.txt file is :" >>>>>>>>>>>>>>>> vbaby2.cloud.eb >>>>>>>>>>>>>>>> vbaby3.cloud.eb >>>>>>>>>>>>>>>> vbaby4.cloud.eb >>>>>>>>>>>>>>>> vbaby5.cloud.eb >>>>>>>>>>>>>>>> vbaby6.cloud.eb >>>>>>>>>>>>>>>> " >>>>>>>>>>>>>>>> Thank you very much ~ >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> 2012-08-09 >>>>>>>>>>>>>>>> xupeng.bupt >>>>>>>>>>>>>>>> 发件人: Artem Ervits >>>>>>>>>>>>>>>> 发送时间: 2012-08-09 22:16:53 >>>>>>>>>>>>>>>> 收件人: [email protected] >>>>>>>>>>>>>>>> 抄送: >>>>>>>>>>>>>>>> 主题: RE: Problem when setting up hadoop cluster step 2 >>>>>>>>>>>>>>>> the installer requires a hosts file which I believe you called >>>>>>>>>>>>>>>> hostdetail. Make sure it's the same file. You also mention a >>>>>>>>>>>>>>>> hosts.txt and host.txt. You only need one file with the names >>>>>>>>>>>>>>>> of all servers. >>>>>>>>>>>>>>>> -----Original Message----- >>>>>>>>>>>>>>>> From: xu peng [mailto:[email protected]] >>>>>>>>>>>>>>>> Sent: Thursday, August 09, 2012 2:02 AM >>>>>>>>>>>>>>>> To: [email protected] >>>>>>>>>>>>>>>> Subject: Problem when setting up hadoop cluster step 2 >>>>>>>>>>>>>>>> Hi everyone : >>>>>>>>>>>>>>>> I am trying to use ambari to set up a hadoop cluster , but i >>>>>>>>>>>>>>>> encounter a problem on step 2. I already set up the >>>>>>>>>>>>>>>> password-less ssh, and i creat a hostdetail.txt file. >>>>>>>>>>>>>>>> The problem is that i found the file >>>>>>>>>>>>>>>> "/var/run/hmc/clusters/EBHadoop/hosts.txt" is empty , no >>>>>>>>>>>>>>>> matter how many times i submit the host.txt file on the >>>>>>>>>>>>>>>> website , and i really don't know why. >>>>>>>>>>>>>>>> { >>>>>>>>>>>>>>>> Here is the log file : [2012:08:09 >>>>>>>>>>>>>>>> 05:17:56][ERROR][sequentialScriptExecutor][sequentialScriptRunner.php:272][]: >>>>>>>>>>>>>>>> Encountered total failure in transaction 100 while running cmd: >>>>>>>>>>>>>>>> /usr/bin/php ./addNodes/findSshableNodes.php with args: >>>>>>>>>>>>>>>> EBHadoop root >>>>>>>>>>>>>>>> 35 100 36 /var/run/hmc/clusters/EBHadoop/hosts.txt >>>>>>>>>>>>>>>> and my host.txt is like this(vbaby1.cloud.eb is the master >>>>>>>>>>>>>>>> node) : >>>>>>>>>>>>>>>> vbaby2.cloud.eb >>>>>>>>>>>>>>>> vbaby3.cloud.eb >>>>>>>>>>>>>>>> vbaby4.cloud.eb >>>>>>>>>>>>>>>> vbaby5.cloud.eb >>>>>>>>>>>>>>>> vbaby6.cloud.eb >>>>>>>>>>>>>>>> } >>>>>>>>>>>>>>>> Can anyone help me and tell me what i am doing wrong ? >>>>>>>>>>>>>>>> Thank you very much ~! >>>>>>>>>>>>>>>> This electronic message is intended to be for the use only of >>>>>>>>>>>>>>>> the named recipient, and may contain information that is >>>>>>>>>>>>>>>> confidential or privileged. If you are not the intended >>>>>>>>>>>>>>>> recipient, you are hereby notified that any disclosure, >>>>>>>>>>>>>>>> copying, distribution or use of the contents of this message >>>>>>>>>>>>>>>> is strictly prohibited. If you have received this message in >>>>>>>>>>>>>>>> error or are not the named recipient, please notify us >>>>>>>>>>>>>>>> immediately by contacting the sender at the electronic mail >>>>>>>>>>>>>>>> address noted above, and delete and destroy all copies of this >>>>>>>>>>>>>>>> message. Thank you. >>>>>>>>>>>>>>>> -------------------- >>>>>>>>>>>>>>>> This electronic message is intended to be for the use only of >>>>>>>>>>>>>>>> the named recipient, and may contain information that is >>>>>>>>>>>>>>>> confidential or privileged. If you are not the intended >>>>>>>>>>>>>>>> recipient, you are hereby notified that any disclosure, >>>>>>>>>>>>>>>> copying, distribution or use of the contents of this message >>>>>>>>>>>>>>>> is strictly prohibited. If you have received this message in >>>>>>>>>>>>>>>> error or are not the named recipient, please notify us >>>>>>>>>>>>>>>> immediately by contacting the sender at the electronic mail >>>>>>>>>>>>>>>> address noted above, and delete and destroy all copies of this >>>>>>>>>>>>>>>> message. Thank you. >>>>>>>>>>>>>>>> -------------------- >>>>>>>>>>>>>>>> This electronic message is intended to be for the use only of >>>>>>>>>>>>>>>> the named recipient, and may contain information that is >>>>>>>>>>>>>>>> confidential or privileged. If you are not the intended >>>>>>>>>>>>>>>> recipient, you are hereby notified that any disclosure, >>>>>>>>>>>>>>>> copying, distribution or use of the contents of this message >>>>>>>>>>>>>>>> is strictly prohibited. If you have received this message in >>>>>>>>>>>>>>>> error or are not the named recipient, please notify us >>>>>>>>>>>>>>>> immediately by contacting the sender at the electronic mail >>>>>>>>>>>>>>>> address noted above, and delete and destroy all copies of this >>>>>>>>>>>>>>>> message. Thank you. >>>>>>>>>>>>>>>> -------------------- >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> This electronic message is intended to be for the use only of >>>>>>>>>>>>>>>> the named recipient, and may contain information that is >>>>>>>>>>>>>>>> confidential or privileged. If you are not the intended >>>>>>>>>>>>>>>> recipient, you are hereby notified that any disclosure, >>>>>>>>>>>>>>>> copying, distribution or use of the contents of this message >>>>>>>>>>>>>>>> is strictly prohibited. If you have received this message in >>>>>>>>>>>>>>>> error or are not the named recipient, please notify us >>>>>>>>>>>>>>>> immediately by contacting the sender at the electronic mail >>>>>>>>>>>>>>>> address noted above, and delete and destroy all copies of this >>>>>>>>>>>>>>>> message. Thank you. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> -------------------- >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> This electronic message is intended to be for the use only of >>>>>>>>>>>>>>>> the named recipient, and may contain information that is >>>>>>>>>>>>>>>> confidential or privileged. If you are not the intended >>>>>>>>>>>>>>>> recipient, you are hereby notified that any disclosure, >>>>>>>>>>>>>>>> copying, distribution or use of the contents of this message >>>>>>>>>>>>>>>> is strictly prohibited. If you have received this message in >>>>>>>>>>>>>>>> error or are not the named recipient, please notify us >>>>>>>>>>>>>>>> immediately by contacting the sender at the electronic mail >>>>>>>>>>>>>>>> address noted above, and delete and destroy all copies of this >>>>>>>>>>>>>>>> message. Thank you. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>> <hmcLog.txt><hostdetails.txt><httpdLog.txt><ssh1.sh><ssh1_result.jpg> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> <DeployError1_2012.8.13.txt><log.rar><DeployingDetails.jpg><UninstallError1_2012.8.13.txt> >>>>> >>> >> <deployError2012.8.17.txt> >
