Kirk,

Reformat is a bit too drastic.

In case of vmware, CS does not cleanly remove its traces and the next re-join 
fails.

However, in that case for vmware at least, all I have to do - is remove:

Network additions by CS
Storage additions by CS (secondary and primary)

I assume something similar can be done, as Bjoern mentioned.

Regards
ilya
> -----Original Message-----
> From: Teipel, Bjoern [mailto:bjoern.tei...@internetbrands.com]
> Sent: Friday, March 08, 2013 1:53 AM
> To: cloudstack-users@incubator.apache.org
> Subject: Re: Domain not found
> 
> I found best works to
> 
> Stop agent
> Remove libvirt pools and interfaces
> Remove /etc/cloud
> Reinstall cloud-agent
> Remove net bridges installed by CS
> 
> After that you have a clean hypervisor
> and can start over
> 
> Bjoern
> 
> On Mar 7, 2013, at 8:45 PM, "Kirk Kosinski" <kirkkosin...@gmail.com> wrote:
> 
> > There are indeed some bugs with similar errors, and some suggested
> > workarounds.
> >
> > https://issues.apache.org/jira/browse/CLOUDSTACK-597
> > Stop CloudStack, stop all KVM agents, start CloudStack, start agents.
> >
> > https://issues.apache.org/jira/browse/CLOUDSTACK-600
> > Disable libvirt-guests.
> >
> > https://issues.apache.org/jira/browse/CLOUDSTACK-1430
> > Check NFS settings.
> >
> > Besides that, I don't think reusing hosts is supported.  I know there
> > are problems with reusing XenServer hosts, so this may also apply to
> > KVM.  You should reformat the host before adding it to a new CloudStack.
> >
> > Best regards,
> > Kirk
> >
> >
> >
> > On 03/07/2013 04:41 PM, Musayev, Ilya wrote:
> >> Have you considered going into JIRA and searching for bug reports that
> contain your issue. If one is found and addressed, consider an upgrade. If not
> found, please file a bug report and include steps to reproduce as well as the
> errors you see.
> >>
> >> We can also look into direct DB changes - but that would be our last resort
> - when nothing else helps.
> >>
> >> I would also search markmail.org for your error, to check if someone else
> may have reported/resolved this issue.
> >>
> >> Also, consider asking on Dev list.
> >>
> >> Regards
> >> Ilya
> >>
> >> Wei Leong <wle...@blackducksoftware.com> wrote:
> >> Hi all,
> >>
> >> I keep getting domain not found warnings in agent.log. I suspect this is
> because the host I am using used to be associated with another
> management server. How do I cleanup the domains in the host so it doesn't
> look for those domains again? Thanks!
> >>
> >> 2013-03-07 09:41:36,625 WARN
> [kvm.resource.LibvirtComputingResource]
> >> (UgentTask-5:null) Unable to get vms
> >> org.libvirt.LibvirtException: Domain not found: no domain with matching
> uuid 'bb897daa-2cab-3546-a188-b8779a2113b4'
> >>        at org.libvirt.ErrorHandler.processError(Unknown Source)
> >>        at org.libvirt.Connect.processError(Unknown Source)
> >>        at org.libvirt.Connect.domainLookupByUUIDString(Unknown Source)
> >>        at org.libvirt.Connect.domainLookupByUUID(Unknown Source)
> >>        at
> com.cloud.hypervisor.kvm.resource.LibvirtComputingResource.getAllVms(Li
> bvirtComputingResource.java:3540)
> >>        at
> com.cloud.hypervisor.kvm.resource.LibvirtComputingResource.sync(LibvirtC
> omputingResource.java:3303)
> >>        at
> com.cloud.hypervisor.kvm.resource.LibvirtComputingResource.getCurrentSt
> atus(LibvirtComputingResource.java:3216)
> >>        at com.cloud.agent.Agent.processOtherTask(Agent.java:590)
> >>        at com.cloud.agent.Agent$ServerHandler.doTask(Agent.java:872)
> >>        at com.cloud.utils.nio.Task.run(Task.java:83)
> >>        at
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
> >>        at
> java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334)
> >>        at java.util.concurrent.FutureTask.run(FutureTask.java:166)
> >>        at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.jav
> a:1110)
> >>        at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.ja
> va:603)
> >>        at java.lang.Thread.run(Thread.java:722)
> >>
> >>
> >>


Reply via email to