Hi Hitesh : I use the default settings of the mount point , but it seems like this path is not a directory(/dev/mapper/hdvg-rootlv/), and i can not execute mkdir -p command on this path. And the hdvg-rootlv is a blocking file (bwrxwrxwrx) . Is there something wrong ?
On Sun, Aug 19, 2012 at 3:38 AM, Hitesh Shah <[email protected]> wrote: > Hi > > Yes - you should all packages from the new repo and none from the old repo. > Most of the packages should be the same but same like hadoop-lzo were > re-factored to work correctly with respect to 32/64-bit installs on RHEL6. > > Regarding the mount points, from a hadoop point of view, the namenode and > datanode dirs are just dirs. From a performance point of view, you want each > dir to be created on a separate mount point to increase disk io bandwidth. > This means that the mount points that you select on the UI should allow > directories to be created. If you have mounted certain kind of filesystems > which you do not wish to use for hadoop ( any tmpfs, nfs mounts etc ), you > should de-select them on the UI and/or use the custom mount point text box as > appropriate. The UI currently does not distinguish valid mount points and > therefore it is up to the user to select correctly. > > -- Hitesh > > > On Aug 18, 2012, at 9:48 AM, xu peng wrote: > >> Hi Hitesh: >> >> Thanks again for your reply. >> >> I solved the dependency problem after updating the hdp repo. >> >> But here comes two new problems : >> 1. I update the new hdp repo , but i create a local repo copy of the >> old hdp repo. And I installed all the rpm package except >> hadoop-lzo-native using the old hdp repo. So it seems like the >> hadoop-lzo-native has some conflct with hadoop-lzo. So , do i have to >> install all the rpm package from the new repo ? >> >> 2. From the error log , i can see a command "mkdir -p /var/.../.. >> (mounting point of hadoop)", but i found the mouting point is not a >> dir , but a blocking file(bwrxwrxwrx). And the execution of this step >> failed. Did i do something wrong ? >> >> I am sorry that this deploy error log is on my company's computer, and >> i will upload it in my next email. >> >> >> Thanks >> -- Xupeng >> >> On Sat, Aug 18, 2012 at 4:43 AM, Hitesh Shah <[email protected]> wrote: >>> Hi again, >>> >>> You are actually hitting a problem caused by some changes in the code which >>> require a modified repo. Unfortunately, I got delayed in modifying the >>> documentation to point to the new repo. >>> >>> Could you try using >>> http://public-repo-1.hortonworks.com/HDP-1.0.1.14/repos/centos5/hdp-release-1.0.1.14-1.el5.noarch.rpm >>> or >>> http://public-repo-1.hortonworks.com/HDP-1.0.1.14/repos/centos6/hdp-release-1.0.1.14-1.el6.noarch.rpm >>> >>> The above should install the yum repo configs to point to the correct repo >>> which will have the lzo packages. >>> >>> -- Hitesh >>> >>> >>> On Aug 16, 2012, at 9:27 PM, xu peng wrote: >>> >>>> Hitesh Shah : >>>> >>>> It is a my my pleasure to fill jira of ambari to help other users . As >>>> a matter of fact, i want to summarize all the problem before i install >>>> ambari cluster successfully. And i will feed back as soon as >>>> possiable. >>>> >>>> Here is another problem i encounter when install hadoop using ambari, >>>> i found a rpm package "hadoop-lzp-native" not in the hdp repo >>>> (baseurl=http://public-repo-1.hortonworks.com/HDP-1.0.13/repos/centos5) >>>> . So i failed againg during deploying step. >>>> >>>> And the attachment is the deploying log , please refer. >>>> >>>> Thanks a lot and look forward to you reply. >>>> >>>> >>>> On Tue, Aug 14, 2012 at 11:35 PM, Hitesh Shah <[email protected]> >>>> wrote: >>>>> Ok - the cert issue is sometimes a result of uninstalling and >>>>> re-installing ambari agents. >>>>> >>>>> The re-install causes ambari agents to regenerate a new certification and >>>>> if the master was bootstrapped earlier, it would still be looking to >>>>> match against old certs. >>>>> >>>>> Stop ambari master and remove ambari-agent rpm from all hosts. >>>>> >>>>> To fix this: >>>>> - on the master, do a puppet cert revoke for all hosts ( >>>>> http://docs.puppetlabs.com/man/cert.html ) >>>>> - you can do a cert list to get all signed or non-signed hosts >>>>> >>>>> On all hosts, delete the following dirs ( if they exist ) : >>>>> - /etc/puppet/ssl >>>>> - /etc/puppet/[master|agent\/ssl/ >>>>> - /var/lib/puppet/ssl/ >>>>> >>>>> >>>>> After doing the above, re-install the ambari agent. >>>>> >>>>> On the ambari master, stop the master. Run the following command: >>>>> >>>>> puppet master --no-daemonize --debug >>>>> >>>>> The above runs in the foreground. The reason to run this is to make sure >>>>> the cert for the master is recreated as we deleted it earlier. >>>>> >>>>> Now, kill the above process running in the foreground and do a service >>>>> ambari start to bring up the UI. >>>>> >>>>> You should be able to bootstrap from this point on. >>>>> >>>>> Would you mind filing a jira and mentioning all the various issues you >>>>> have come across and how you solved them. We can use that to create an >>>>> FAQ for other users. >>>>> >>>>> thanks >>>>> -- Hitesh >>>>> >>>>> >>>>> On Aug 14, 2012, at 1:55 AM, xu peng wrote: >>>>> >>>>>> Hi Hitesh : >>>>>> >>>>>> Thanks a lot for your reply. >>>>>> >>>>>> 1. I did a puppet kick --ping to the client from my ambari master , >>>>>> all the five nodes failed with the same log (Triggering >>>>>> vbaby2.cloud.eb >>>>>> Host vbaby2.cloud.eb failed: certificate verify failed. This is often >>>>>> because the time is out of sync on the server or client >>>>>> vbaby2.cloud.eb finished with exit code 2) >>>>>> >>>>>> I manually run "service ambari-agent start" , is that necessary ? How >>>>>> can i fix these problem ? >>>>>> >>>>>> 2. As you suggest , I run the yum command manually. And found that the >>>>>> installation missed some dependecy - php-gd. And i have to update my >>>>>> yum repo. >>>>>> >>>>>> >>>>>> >>>>>> On Tue, Aug 14, 2012 at 1:01 AM, Hitesh Shah <[email protected]> >>>>>> wrote: >>>>>>> Based on your deploy error log: >>>>>>> >>>>>>> "3": { >>>>>>> "nodeReport": { >>>>>>> "PUPPET_KICK_FAILED": [], >>>>>>> "PUPPET_OPERATION_FAILED": [ >>>>>>> "vbaby3.cloud.eb", >>>>>>> "vbaby5.cloud.eb", >>>>>>> "vbaby4.cloud.eb", >>>>>>> "vbaby2.cloud.eb", >>>>>>> "vbaby6.cloud.eb", >>>>>>> "vbaby1.cloud.eb" >>>>>>> ], >>>>>>> "PUPPET_OPERATION_TIMEDOUT": [ >>>>>>> "vbaby5.cloud.eb", >>>>>>> "vbaby4.cloud.eb", >>>>>>> "vbaby2.cloud.eb", >>>>>>> "vbaby6.cloud.eb", >>>>>>> "vbaby1.cloud.eb" >>>>>>> ], >>>>>>> >>>>>>> 5 nodes timed out which means the puppet agent is not running on them >>>>>>> or they cannot communicate with the master. Trying doing a puppet kick >>>>>>> --ping to them from the master. >>>>>>> >>>>>>> For the one which failed, it failed at >>>>>>> >>>>>>> "\"Mon Aug 13 11:54:17 +0800 2012 >>>>>>> /Stage[1]/Hdp::Pre_install_pkgs/Hdp::Exec[yum install >>>>>>> $pre_installed_pkgs]/Exec[yum install $pre_installed_pkgs]/returns >>>>>>> (err): change from notrun to 0 failed: yum install -y hadoop >>>>>>> hadoop-libhdfs hadoop-native hadoop-pipes hadoop-sbin hadoop-lzo hadoop >>>>>>> hadoop-libhdfs hadoop-native hadoop-pipes hadoop-sbin hadoop-lzo >>>>>>> hdp_mon_dashboard ganglia-gmond-3.2.0 gweb hdp_mon_ganglia_addons >>>>>>> snappy snappy-devel returned 1 instead of one of [0] at >>>>>>> /etc/puppet/agent/modules/hdp/manifests/init.pp:265\"", >>>>>>> >>>>>>> It seems like yum install failed on the host. Try running the command >>>>>>> manually and see what the error is. >>>>>>> >>>>>>> -- Hitesh >>>>>>> >>>>>>> >>>>>>> >>>>>>> On Aug 13, 2012, at 2:28 AM, xu peng wrote: >>>>>>> >>>>>>>> Hi Hitesh : >>>>>>>> >>>>>>>> It's me again. >>>>>>>> >>>>>>>> Followed you advice , I reinstalled the ambari server. But deploying >>>>>>>> cluster and uninstall cluster failed again. I really don't know why. >>>>>>>> >>>>>>>> I supplied a attachment which contains the logs of all the nodes in >>>>>>>> my cluster (/var/log/puppet_*.log , /var/log/puppet/*.log , >>>>>>>> /var/log/yum.log, /var/log/hmc/hmc.log). And vbaby3.cloud.eb is the >>>>>>>> ambari server. Please refer. >>>>>>>> >>>>>>>> Attachment DeployError and UninstallError is the log supplied by the >>>>>>>> website of ambari when failing. And attachment DeployingDetails.jpg is >>>>>>>> the deploy details of my cluster. Please refer. >>>>>>>> >>>>>>>> >>>>>>>> Thanks again for your patience ! And look forward to your reply. >>>>>>>> >>>>>>>> Xupeng >>>>>>>> >>>>>>>> On Sat, Aug 11, 2012 at 10:56 PM, Hitesh Shah <[email protected]> >>>>>>>> wrote: >>>>>>>>> For uninstall failures, you will need to do a couple of things. >>>>>>>>> Depending on where the uninstall failed, you may have to manually do >>>>>>>>> a killall java on all the nodes to kill any missed processes. If you >>>>>>>>> want to start with a complete clean install, you should also delete >>>>>>>>> the hadoop dir in the mount points you selected during the previous >>>>>>>>> install so that the new fresh install does not face errors when it >>>>>>>>> tries to re-format hdfs. >>>>>>>>> >>>>>>>>> After that, simply, uinstall and re-install ambari rpm and that >>>>>>>>> should allow you to re-create a fresh cluster. >>>>>>>>> >>>>>>>>> -- Hitesh >>>>>>>>> >>>>>>>>> On Aug 11, 2012, at 2:34 AM, xu peng wrote: >>>>>>>>> >>>>>>>>>> Hi Hitesh : >>>>>>>>>> >>>>>>>>>> Thanks a lot for your reply. >>>>>>>>>> >>>>>>>>>> I solved this problem , it is silly mistake. Someone has changed the >>>>>>>>>> owner of "/" dir , and according to the errorlog , pdsh need root to >>>>>>>>>> proceed. >>>>>>>>>> >>>>>>>>>> After changing the owner of "/" to root , problem solved. Thank you >>>>>>>>>> again for you reply. >>>>>>>>>> >>>>>>>>>> I have another question. I had a uninstall failure , and there is no >>>>>>>>>> button on the website for me to rollback and i don't know what to do >>>>>>>>>> about that. What should i do now to reinstall hadoop ? >>>>>>>>>> >>>>>>>>>> Thanks >>>>>>>>>> >>>>>>>>>> On Fri, Aug 10, 2012 at 10:55 PM, Hitesh Shah >>>>>>>>>> <[email protected]> wrote: >>>>>>>>>>> Hi >>>>>>>>>>> >>>>>>>>>>> Currently, the ambari installer requires everything to be run as >>>>>>>>>>> root. It does not detect that the user is not root and use sudo >>>>>>>>>>> either on the master or on the agent nodes. >>>>>>>>>>> Furthermore, it seems like it is failing when trying to use pdsh to >>>>>>>>>>> make remote calls to the host list that you passed in due to the >>>>>>>>>>> errors mentioned in your script. This could be due to how it was >>>>>>>>>>> installed but I am not sure. >>>>>>>>>>> >>>>>>>>>>> Could you switch to become root and run any simple command on all >>>>>>>>>>> hosts using pdsh? If you want to reference exactly how ambari uses >>>>>>>>>>> pdsh, you can look into /usr/share/hmc/php/frontend/commandUtils.php >>>>>>>>>>> >>>>>>>>>>> thanks >>>>>>>>>>> -- Hitesh >>>>>>>>>>> >>>>>>>>>>> On Aug 9, 2012, at 9:04 PM, xu peng wrote: >>>>>>>>>>> >>>>>>>>>>>> According to the error log , is there something wrong with my >>>>>>>>>>>> account ? >>>>>>>>>>>> >>>>>>>>>>>> I installed all the dependency module and ambari with the user >>>>>>>>>>>> "ambari" instead of root. I added user "ambari" to /etc/sudofilers >>>>>>>>>>>> with no passwd. >>>>>>>>>>>> >>>>>>>>>>>> On Fri, Aug 10, 2012 at 11:49 AM, xu peng <[email protected]> >>>>>>>>>>>> wrote: >>>>>>>>>>>>> There is no 100.log.file in /var/log/hmc dir, but only 55.log >>>>>>>>>>>>> file (55 >>>>>>>>>>>>> is the biggest version num). >>>>>>>>>>>>> >>>>>>>>>>>>> The content of 55.log is : >>>>>>>>>>>>> pdsh@vbaby1: module path "/usr/lib64/pdsh" insecure. >>>>>>>>>>>>> pdsh@vbaby1: "/": Owner not root, current uid, or pdsh executable >>>>>>>>>>>>> owner >>>>>>>>>>>>> pdsh@vbaby1: Couldn't load any pdsh modules >>>>>>>>>>>>> >>>>>>>>>>>>> Thanks ~ >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> On Fri, Aug 10, 2012 at 11:36 AM, Hitesh Shah >>>>>>>>>>>>> <[email protected]> wrote: >>>>>>>>>>>>>> Sorry - my mistake. The last txn mentioned is 100 so please look >>>>>>>>>>>>>> for the 100.log file. >>>>>>>>>>>>>> >>>>>>>>>>>>>> -- Hitesh >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> On Aug 9, 2012, at 8:34 PM, Hitesh Shah wrote: >>>>>>>>>>>>>> >>>>>>>>>>>>>>> Thanks - will take a look and get back to you. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Could you also look at /var/log/hmc/hmc.txn.55.log and see if >>>>>>>>>>>>>>> there are any errors in it? >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> -- Hitesh. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> On Aug 9, 2012, at 8:00 PM, xu peng wrote: >>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Hi Hitesh : >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Thanks a lot for your replying. I have done all your >>>>>>>>>>>>>>>> suggestions in my >>>>>>>>>>>>>>>> ambari server , and the result is as below. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> 1. I can confirm that the hosts.txt file is empty after i >>>>>>>>>>>>>>>> failed at >>>>>>>>>>>>>>>> the step finding reachable nodes. >>>>>>>>>>>>>>>> 2. I tried make hostdetails file in win7 and redhat , it both >>>>>>>>>>>>>>>> failed.(Please see the attachment, my hostdetails file) >>>>>>>>>>>>>>>> 3. I removed the logging re-direct and run the .sh script .It >>>>>>>>>>>>>>>> seems >>>>>>>>>>>>>>>> like the script works well , it print the hostname in console >>>>>>>>>>>>>>>> and >>>>>>>>>>>>>>>> generate a file (content is "0") in the same dir. (Please see >>>>>>>>>>>>>>>> the >>>>>>>>>>>>>>>> attachment , the result and my .sh script ) >>>>>>>>>>>>>>>> 4. I attached the hmc.log and error_log too. Hope this helps ~ >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Thanks ~ >>>>>>>>>>>>>>>> Xupeng >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> On Fri, Aug 10, 2012 at 12:24 AM, Hitesh Shah >>>>>>>>>>>>>>>> <[email protected]> wrote: >>>>>>>>>>>>>>>>> Xupeng, can you confirm that the hosts.txt file at >>>>>>>>>>>>>>>>> /var/run/hmc/clusters/EBHadoop/hosts.txt is empty? >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Also, can you ensure that the hostdetails file that you >>>>>>>>>>>>>>>>> upload does not have any special characters that may be >>>>>>>>>>>>>>>>> creating problems for the parsing layer? >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> In the same dir, there should be an ssh.sh script. Can you >>>>>>>>>>>>>>>>> create a copy of it, edit to remove the logging re-directs to >>>>>>>>>>>>>>>>> files and run the script manually from command-line ( it >>>>>>>>>>>>>>>>> takes in a hostname as the argument ) ? The output of that >>>>>>>>>>>>>>>>> should show you as to what is going wrong. >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Also, please look at /var/log/hmc/hmc.log and httpd/error_log >>>>>>>>>>>>>>>>> to see if there are any errors being logged which may shed >>>>>>>>>>>>>>>>> more light on the issue. >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> thanks >>>>>>>>>>>>>>>>> -- Hitesh >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> On Aug 9, 2012, at 9:11 AM, Artem Ervits wrote: >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> Which file are you supplying in the step? Hostdetail.txt or >>>>>>>>>>>>>>>>>> hosts? >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> From: xupeng.bupt [mailto:[email protected]] >>>>>>>>>>>>>>>>>> Sent: Thursday, August 09, 2012 11:33 AM >>>>>>>>>>>>>>>>>> To: ambari-user >>>>>>>>>>>>>>>>>> Subject: Re: RE: Problem when setting up hadoop cluster step >>>>>>>>>>>>>>>>>> 2 >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> Thank you for your replying ~ >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> I made only one hostdetail.txt file which contains the names >>>>>>>>>>>>>>>>>> of all servers. And i submit this file on the website , but >>>>>>>>>>>>>>>>>> i still have the same problem. I failed at the step of >>>>>>>>>>>>>>>>>> finding reachable nodes. >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> The error log is : " >>>>>>>>>>>>>>>>>> [ERROR][sequentialScriptExecutor][sequentialScriptRunner.php:272][]: >>>>>>>>>>>>>>>>>> Encountered total failure in transaction 100 while running >>>>>>>>>>>>>>>>>> cmd: >>>>>>>>>>>>>>>>>> /usr/bin/php ./addNodes/findSshableNodes.php with args: >>>>>>>>>>>>>>>>>> EBHadoop root >>>>>>>>>>>>>>>>>> 35 100 36 /var/run/hmc/clusters/EBHadoop/hosts.txt >>>>>>>>>>>>>>>>>> " >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> And my hostdetail.txt file is :" >>>>>>>>>>>>>>>>>> vbaby2.cloud.eb >>>>>>>>>>>>>>>>>> vbaby3.cloud.eb >>>>>>>>>>>>>>>>>> vbaby4.cloud.eb >>>>>>>>>>>>>>>>>> vbaby5.cloud.eb >>>>>>>>>>>>>>>>>> vbaby6.cloud.eb >>>>>>>>>>>>>>>>>> " >>>>>>>>>>>>>>>>>> Thank you very much ~ >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> 2012-08-09 >>>>>>>>>>>>>>>>>> xupeng.bupt >>>>>>>>>>>>>>>>>> 发件人: Artem Ervits >>>>>>>>>>>>>>>>>> 发送时间: 2012-08-09 22:16:53 >>>>>>>>>>>>>>>>>> 收件人: [email protected] >>>>>>>>>>>>>>>>>> 抄送: >>>>>>>>>>>>>>>>>> 主题: RE: Problem when setting up hadoop cluster step 2 >>>>>>>>>>>>>>>>>> the installer requires a hosts file which I believe you >>>>>>>>>>>>>>>>>> called hostdetail. Make sure it's the same file. You also >>>>>>>>>>>>>>>>>> mention a hosts.txt and host.txt. You only need one file >>>>>>>>>>>>>>>>>> with the names of all servers. >>>>>>>>>>>>>>>>>> -----Original Message----- >>>>>>>>>>>>>>>>>> From: xu peng [mailto:[email protected]] >>>>>>>>>>>>>>>>>> Sent: Thursday, August 09, 2012 2:02 AM >>>>>>>>>>>>>>>>>> To: [email protected] >>>>>>>>>>>>>>>>>> Subject: Problem when setting up hadoop cluster step 2 >>>>>>>>>>>>>>>>>> Hi everyone : >>>>>>>>>>>>>>>>>> I am trying to use ambari to set up a hadoop cluster , but i >>>>>>>>>>>>>>>>>> encounter a problem on step 2. I already set up the >>>>>>>>>>>>>>>>>> password-less ssh, and i creat a hostdetail.txt file. >>>>>>>>>>>>>>>>>> The problem is that i found the file >>>>>>>>>>>>>>>>>> "/var/run/hmc/clusters/EBHadoop/hosts.txt" is empty , no >>>>>>>>>>>>>>>>>> matter how many times i submit the host.txt file on the >>>>>>>>>>>>>>>>>> website , and i really don't know why. >>>>>>>>>>>>>>>>>> { >>>>>>>>>>>>>>>>>> Here is the log file : [2012:08:09 >>>>>>>>>>>>>>>>>> 05:17:56][ERROR][sequentialScriptExecutor][sequentialScriptRunner.php:272][]: >>>>>>>>>>>>>>>>>> Encountered total failure in transaction 100 while running >>>>>>>>>>>>>>>>>> cmd: >>>>>>>>>>>>>>>>>> /usr/bin/php ./addNodes/findSshableNodes.php with args: >>>>>>>>>>>>>>>>>> EBHadoop root >>>>>>>>>>>>>>>>>> 35 100 36 /var/run/hmc/clusters/EBHadoop/hosts.txt >>>>>>>>>>>>>>>>>> and my host.txt is like this(vbaby1.cloud.eb is the master >>>>>>>>>>>>>>>>>> node) : >>>>>>>>>>>>>>>>>> vbaby2.cloud.eb >>>>>>>>>>>>>>>>>> vbaby3.cloud.eb >>>>>>>>>>>>>>>>>> vbaby4.cloud.eb >>>>>>>>>>>>>>>>>> vbaby5.cloud.eb >>>>>>>>>>>>>>>>>> vbaby6.cloud.eb >>>>>>>>>>>>>>>>>> } >>>>>>>>>>>>>>>>>> Can anyone help me and tell me what i am doing wrong ? >>>>>>>>>>>>>>>>>> Thank you very much ~! >>>>>>>>>>>>>>>>>> This electronic message is intended to be for the use only >>>>>>>>>>>>>>>>>> of the named recipient, and may contain information that is >>>>>>>>>>>>>>>>>> confidential or privileged. If you are not the intended >>>>>>>>>>>>>>>>>> recipient, you are hereby notified that any disclosure, >>>>>>>>>>>>>>>>>> copying, distribution or use of the contents of this message >>>>>>>>>>>>>>>>>> is strictly prohibited. If you have received this message in >>>>>>>>>>>>>>>>>> error or are not the named recipient, please notify us >>>>>>>>>>>>>>>>>> immediately by contacting the sender at the electronic mail >>>>>>>>>>>>>>>>>> address noted above, and delete and destroy all copies of >>>>>>>>>>>>>>>>>> this message. Thank you. >>>>>>>>>>>>>>>>>> -------------------- >>>>>>>>>>>>>>>>>> This electronic message is intended to be for the use only >>>>>>>>>>>>>>>>>> of the named recipient, and may contain information that is >>>>>>>>>>>>>>>>>> confidential or privileged. If you are not the intended >>>>>>>>>>>>>>>>>> recipient, you are hereby notified that any disclosure, >>>>>>>>>>>>>>>>>> copying, distribution or use of the contents of this message >>>>>>>>>>>>>>>>>> is strictly prohibited. If you have received this message >>>>>>>>>>>>>>>>>> in error or are not the named recipient, please notify us >>>>>>>>>>>>>>>>>> immediately by contacting the sender at the electronic mail >>>>>>>>>>>>>>>>>> address noted above, and delete and destroy all copies of >>>>>>>>>>>>>>>>>> this message. Thank you. >>>>>>>>>>>>>>>>>> -------------------- >>>>>>>>>>>>>>>>>> This electronic message is intended to be for the use only >>>>>>>>>>>>>>>>>> of the named recipient, and may contain information that is >>>>>>>>>>>>>>>>>> confidential or privileged. If you are not the intended >>>>>>>>>>>>>>>>>> recipient, you are hereby notified that any disclosure, >>>>>>>>>>>>>>>>>> copying, distribution or use of the contents of this message >>>>>>>>>>>>>>>>>> is strictly prohibited. If you have received this message >>>>>>>>>>>>>>>>>> in error or are not the named recipient, please notify us >>>>>>>>>>>>>>>>>> immediately by contacting the sender at the electronic mail >>>>>>>>>>>>>>>>>> address noted above, and delete and destroy all copies of >>>>>>>>>>>>>>>>>> this message. Thank you. >>>>>>>>>>>>>>>>>> -------------------- >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> This electronic message is intended to be for the use only >>>>>>>>>>>>>>>>>> of the named recipient, and may contain information that is >>>>>>>>>>>>>>>>>> confidential or privileged. If you are not the intended >>>>>>>>>>>>>>>>>> recipient, you are hereby notified that any disclosure, >>>>>>>>>>>>>>>>>> copying, distribution or use of the contents of this message >>>>>>>>>>>>>>>>>> is strictly prohibited. If you have received this message >>>>>>>>>>>>>>>>>> in error or are not the named recipient, please notify us >>>>>>>>>>>>>>>>>> immediately by contacting the sender at the electronic mail >>>>>>>>>>>>>>>>>> address noted above, and delete and destroy all copies of >>>>>>>>>>>>>>>>>> this message. Thank you. >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> -------------------- >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> This electronic message is intended to be for the use only >>>>>>>>>>>>>>>>>> of the named recipient, and may contain information that is >>>>>>>>>>>>>>>>>> confidential or privileged. If you are not the intended >>>>>>>>>>>>>>>>>> recipient, you are hereby notified that any disclosure, >>>>>>>>>>>>>>>>>> copying, distribution or use of the contents of this message >>>>>>>>>>>>>>>>>> is strictly prohibited. If you have received this message >>>>>>>>>>>>>>>>>> in error or are not the named recipient, please notify us >>>>>>>>>>>>>>>>>> immediately by contacting the sender at the electronic mail >>>>>>>>>>>>>>>>>> address noted above, and delete and destroy all copies of >>>>>>>>>>>>>>>>>> this message. Thank you. >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> <hmcLog.txt><hostdetails.txt><httpdLog.txt><ssh1.sh><ssh1_result.jpg> >>>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>>> <DeployError1_2012.8.13.txt><log.rar><DeployingDetails.jpg><UninstallError1_2012.8.13.txt> >>>>>>> >>>>> >>>> <deployError2012.8.17.txt> >>> >
