So this morning I reconnected all hosts. I also disabled my two hosts that need to reboot and powered on a VM and now I am getting a Insufficient Resources.
Whats odd is that Host Allocator returning 0 suitable hosts? 2017-06-21 08:43:53,695 DEBUG [c.c.v.VirtualMachineManagerImpl] (Work-Job-Executor-71:ctx-f01a90b9 job-146764/job-146768 ctx-66c78348) Root volume is ready, need to place VM in volume's cluster 2017-06-21 08:43:53,695 DEBUG [c.c.v.VirtualMachineManagerImpl] (Work-Job-Executor-71:ctx-f01a90b9 job-146764/job-146768 ctx-66c78348) Vol[537|vm=446|ROOT] is READY, changing deployment plan to use this pool's dcId: 1 , podId: 1 , and clusterId: 1 2017-06-21 08:43:53,702 DEBUG [c.c.d.DeploymentPlanningManagerImpl] (Work-Job-Executor-71:ctx-f01a90b9 job-146764/job-146768 ctx-66c78348) Deploy avoids pods: [], clusters: [], hosts: [17, 1, 19, 15] 2017-06-21 08:43:53,703 DEBUG [c.c.d.DeploymentPlanningManagerImpl] (Work-Job-Executor-71:ctx-f01a90b9 job-146764/job-146768 ctx-66c78348) DeploymentPlanner allocation algorithm: com.cloud.deploy.UserDispersingPlanner@4cafa203 2017-06-21 08:43:53,703 DEBUG [c.c.d.DeploymentPlanningManagerImpl] (Work-Job-Executor-71:ctx-f01a90b9 job-146764/job-146768 ctx-66c78348) Trying to allocate a host and storage pools from dc:1, pod:1,cluster:1, requested cpu: 8000, requested ram: 12884901888 2017-06-21 08:43:53,703 DEBUG [c.c.d.DeploymentPlanningManagerImpl] (Work-Job-Executor-71:ctx-f01a90b9 job-146764/job-146768 ctx-66c78348) Is ROOT volume READY (pool already allocated)?: Yes 2017-06-21 08:43:53,703 DEBUG [c.c.d.DeploymentPlanningManagerImpl] (Work-Job-Executor-71:ctx-f01a90b9 job-146764/job-146768 ctx-66c78348) This VM has last host_id specified, trying to choose the same host: 1 2017-06-21 08:43:53,704 DEBUG [c.c.d.DeploymentPlanningManagerImpl] (Work-Job-Executor-71:ctx-f01a90b9 job-146764/job-146768 ctx-66c78348) The last host of this VM is in avoid set 2017-06-21 08:43:53,704 DEBUG [c.c.d.DeploymentPlanningManagerImpl] (Work-Job-Executor-71:ctx-f01a90b9 job-146764/job-146768 ctx-66c78348) Cannot choose the last host to deploy this VM 2017-06-21 08:43:53,704 DEBUG [c.c.d.FirstFitPlanner] (Work-Job-Executor-71:ctx-f01a90b9 job-146764/job-146768 ctx-66c78348) Searching resources only under specified Cluster: 1 2017-06-21 08:43:53,714 DEBUG [c.c.d.DeploymentPlanningManagerImpl] (Work-Job-Executor-71:ctx-f01a90b9 job-146764/job-146768 ctx-66c78348) Checking resources in Cluster: 1 under Pod: 1 2017-06-21 08:43:53,714 DEBUG [c.c.a.m.a.i.FirstFitAllocator] (Work-Job-Executor-71:ctx-f01a90b9 job-146764/job-146768 ctx-66c78348 FirstFitRoutingAllocator) Looking for hosts in dc: 1 pod:1 cluster:1 2017-06-21 08:43:53,718 DEBUG [c.c.a.m.a.i.FirstFitAllocator] (Work-Job-Executor-71:ctx-f01a90b9 job-146764/job-146768 ctx-66c78348 FirstFitRoutingAllocator) List of hosts in ascending order of number of VMs: [15, 17, 19, 1, 16, 18] 2017-06-21 08:43:53,718 DEBUG [c.c.a.m.a.i.FirstFitAllocator] (Work-Job-Executor-71:ctx-f01a90b9 job-146764/job-146768 ctx-66c78348 FirstFitRoutingAllocator) FirstFitAllocator has 4 hosts to check for allocation: [Host[-15-Routing], Host[-17-Routing], Host[-19-Routing], Host[-1-Routing]] 2017-06-21 08:43:53,727 DEBUG [c.c.a.m.a.i.FirstFitAllocator] (Work-Job-Executor-71:ctx-f01a90b9 job-146764/job-146768 ctx-66c78348 FirstFitRoutingAllocator) Found 4 hosts for allocation after prioritization: [Host[-15-Routing], Host[-17-Routing], Host[-19-Routing], Host[-1-Routing]] 2017-06-21 08:43:53,727 DEBUG [c.c.a.m.a.i.FirstFitAllocator] (Work-Job-Executor-71:ctx-f01a90b9 job-146764/job-146768 ctx-66c78348 FirstFitRoutingAllocator) Looking for speed=8000Mhz, Ram=12288 2017-06-21 08:43:53,727 DEBUG [c.c.a.m.a.i.FirstFitAllocator] (Work-Job-Executor-71:ctx-f01a90b9 job-146764/job-146768 ctx-66c78348 FirstFitRoutingAllocator) Host name: Flex-Xen3.flexhost.local, hostId: 15 is in avoid set, skipping this and trying other available hosts 2017-06-21 08:43:53,727 DEBUG [c.c.a.m.a.i.FirstFitAllocator] (Work-Job-Executor-71:ctx-f01a90b9 job-146764/job-146768 ctx-66c78348 FirstFitRoutingAllocator) Host name: Flex-Xen4.flexhost.local, hostId: 17 is in avoid set, skipping this and trying other available hosts 2017-06-21 08:43:53,727 DEBUG [c.c.a.m.a.i.FirstFitAllocator] (Work-Job-Executor-71:ctx-f01a90b9 job-146764/job-146768 ctx-66c78348 FirstFitRoutingAllocator) Host name: Flex-Xen1.flexhost.local, hostId: 19 is in avoid set, skipping this and trying other available hosts 2017-06-21 08:43:53,727 DEBUG [c.c.a.m.a.i.FirstFitAllocator] (Work-Job-Executor-71:ctx-f01a90b9 job-146764/job-146768 ctx-66c78348 FirstFitRoutingAllocator) Host name: Flex-Xen2.flexhost.local, hostId: 1 is in avoid set, skipping this and trying other available hosts 2017-06-21 08:43:53,727 DEBUG [c.c.a.m.a.i.FirstFitAllocator] (Work-Job-Executor-71:ctx-f01a90b9 job-146764/job-146768 ctx-66c78348 FirstFitRoutingAllocator) Host Allocator returning 0 suitable hosts 2017-06-21 08:43:53,727 DEBUG [c.c.d.DeploymentPlanningManagerImpl] (Work-Job-Executor-71:ctx-f01a90b9 job-146764/job-146768 ctx-66c78348) No suitable hosts found 2017-06-21 08:43:53,727 DEBUG [c.c.d.DeploymentPlanningManagerImpl] (Work-Job-Executor-71:ctx-f01a90b9 job-146764/job-146768 ctx-66c78348) No suitable hosts found under this Cluster: 1 2017-06-21 08:43:53,728 DEBUG [c.c.d.DeploymentPlanningManagerImpl] (Work-Job-Executor-71:ctx-f01a90b9 job-146764/job-146768 ctx-66c78348) Could not find suitable Deployment Destination for this VM under any clusters, returning. 2017-06-21 08:43:53,728 DEBUG [c.c.d.FirstFitPlanner] (Work-Job-Executor-71:ctx-f01a90b9 job-146764/job-146768 ctx-66c78348) Searching resources only under specified Cluster: 1 2017-06-21 08:43:53,729 DEBUG [c.c.d.FirstFitPlanner] (Work-Job-Executor-71:ctx-f01a90b9 job-146764/job-146768 ctx-66c78348) The specified cluster is in avoid set, returning. 2017-06-21 08:43:53,736 DEBUG [c.c.d.DeploymentPlanningManagerImpl] (Work-Job-Executor-71:ctx-f01a90b9 job-146764/job-146768 ctx-66c78348) Deploy avoids pods: [], clusters: [1], hosts: [17, 1, 19, 15] 2017-06-21 08:43:53,737 DEBUG [c.c.d.DeploymentPlanningManagerImpl] (Work-Job-Executor-71:ctx-f01a90b9 job-146764/job-146768 ctx-66c78348) DeploymentPlanner allocation algorithm: com.cloud.deploy.UserDispersingPlanner@4cafa203 2017-06-21 08:43:53,737 DEBUG [c.c.d.DeploymentPlanningManagerImpl] (Work-Job-Executor-71:ctx-f01a90b9 job-146764/job-146768 ctx-66c78348) Trying to allocate a host and storage pools from dc:1, pod:1,cluster:null, requested cpu: 8000, requested ram: 12884901888 2017-06-21 08:43:53,737 DEBUG [c.c.d.DeploymentPlanningManagerImpl] (Work-Job-Executor-71:ctx-f01a90b9 job-146764/job-146768 ctx-66c78348) Is ROOT volume READY (pool already allocated)?: No 2017-06-21 08:43:53,737 DEBUG [c.c.d.DeploymentPlanningManagerImpl] (Work-Job-Executor-71:ctx-f01a90b9 job-146764/job-146768 ctx-66c78348) This VM has last host_id specified, trying to choose the same host: 1 2017-06-21 08:43:53,739 DEBUG [c.c.d.DeploymentPlanningManagerImpl] (Work-Job-Executor-71:ctx-f01a90b9 job-146764/job-146768 ctx-66c78348) The last host of this VM is in avoid set All oddities. So I did a force reconnect on all 6 hosts and enabled the two hosts that were pending updates. Jeremy -----Original Message----- From: Jeremy Peterson [mailto:jpeter...@acentek.net] Sent: Tuesday, June 20, 2017 12:33 PM To: users@cloudstack.apache.org; S. Brüseke - proIO GmbH <s.brues...@proio.com> Subject: RE: Recreating SystemVM's Ok so my issues have not gone away. I have two hosts that have not rebooted yet tonight I will be maintenancing those hosts out and migrating vm's away from those hosts and then performing a reboot of the host and installing a couple xenserver updates. One thing is I am not getting the CANNOT ATTACH NETWORK error anymore which is cool but. https://drive.google.com/open?id=0B5IXhrpPAT9qQ0FFUmRyRjN4NlE Take a look at creation of VM 20685 2017-06-20 12:15:48,083 DEBUG [c.c.d.DeploymentPlanningManagerImpl] (Work-Job-Executor-82:ctx-c39fa1f8 job-1042/job-138603 ctx-c17ce6fc) Found a potential host id: 1 name: Flex-Xen2.flexhost.local and associated storage pools for this VM 2017-06-20 12:15:48,084 DEBUG [c.c.d.DeploymentPlanningManagerImpl] (Work-Job-Executor-82:ctx-c39fa1f8 job-1042/job-138603 ctx-c17ce6fc) Returning Deployment Destination: Dest[Zone(Id)-Pod(Id)-Cluster(Id)-Host(Id)-Storage(Volume(Id|Type-->Pool(Id))] : Dest[Zone(1)-Pod(1)-Cluster(1)-Host(1)-Storage(Volume(25604|ROOT-->Pool(5))] 2017-06-20 12:15:48,084 DEBUG [c.c.v.VirtualMachineManagerImpl] (Work-Job-Executor-82:ctx-c39fa1f8 job-1042/job-138603 ctx-c17ce6fc) Deployment found - P0=VM[SecondaryStorageVm|s-20685-VM], P0=Dest[Zone(Id)-Pod(Id)-Cluster(Id)-Host(Id)-Storage(Volume(Id|Type-->Pool(Id))] : Dest[Zone(1)-Pod(1)-Cluster(1)-Host(1)-Storage(Volume(25604|ROOT-->Pool(5))] So it found a host and storage pool Networks were already created on line 482-484 But then look it fails on create volume UUID is invalid??? 2017-06-20 12:15:48,262 DEBUG [c.c.v.VirtualMachineManagerImpl] (Work-Job-Executor-88:ctx-c51dafa0 job-342/job-138604 ctx-75edebb0) VM is being created in podId: 1 2017-06-20 12:15:48,264 DEBUG [o.a.c.e.o.NetworkOrchestrator] (Work-Job-Executor-88:ctx-c51dafa0 job-342/job-138604 ctx-75edebb0) Network id=200 is already implemented 2017-06-20 12:15:48,269 DEBUG [c.c.n.g.PodBasedNetworkGuru] (Work-Job-Executor-82:ctx-c39fa1f8 job-1042/job-138603 ctx-c17ce6fc) Allocated a nic NicProfile[81905-20685-0493941d-d193-4325-84bc-d325a8900332-10.90.2.207-null for VM[SecondaryStorageVm|s-20685-VM] 2017-06-20 12:15:48,280 DEBUG [o.a.c.e.o.NetworkOrchestrator] (Work-Job-Executor-82:ctx-c39fa1f8 job-1042/job-138603 ctx-c17ce6fc) Network id=203 is already implemented 2017-06-20 12:15:48,290 DEBUG [o.a.c.e.o.NetworkOrchestrator] (Work-Job-Executor-88:ctx-c51dafa0 job-342/job-138604 ctx-75edebb0) Network id=202 is already implemented 2017-06-20 12:15:48,316 DEBUG [c.c.n.g.StorageNetworkGuru] (Work-Job-Executor-82:ctx-c39fa1f8 job-1042/job-138603 ctx-c17ce6fc) Allocated a storage nic NicProfile[81906-20685-0493941d-d193-4325-84bc-d325a8900332-10.83.2.205-null for VM[SecondaryStorageVm|s-20685-VM] 2017-06-20 12:15:48,336 DEBUG [o.a.c.e.o.VolumeOrchestrator] (Work-Job-Executor-82:ctx-c39fa1f8 job-1042/job-138603 ctx-c17ce6fc) Checking if we need to prepare 1 volumes for VM[SecondaryStorageVm|s-20685-VM] 2017-06-20 12:15:48,342 DEBUG [o.a.c.s.i.TemplateDataFactoryImpl] (Work-Job-Executor-82:ctx-c39fa1f8 job-1042/job-138603 ctx-c17ce6fc) template 1 is already in store:5, type:Image 2017-06-20 12:15:48,344 DEBUG [o.a.c.s.i.TemplateDataFactoryImpl] (Work-Job-Executor-82:ctx-c39fa1f8 job-1042/job-138603 ctx-c17ce6fc) template 1 is already in store:5, type:Primary 2017-06-20 12:15:48,346 DEBUG [o.a.c.e.o.NetworkOrchestrator] (Work-Job-Executor-88:ctx-c51dafa0 job-342/job-138604 ctx-75edebb0) Network id=201 is already implemented 2017-06-20 12:15:48,372 DEBUG [c.c.d.d.DataCenterIpAddressDaoImpl] (Work-Job-Executor-88:ctx-c51dafa0 job-342/job-138604 ctx-75edebb0) Releasing ip address for instance=49817 2017-06-20 12:15:48,381 DEBUG [o.a.c.s.m.AncientDataMotionStrategy] (Work-Job-Executor-82:ctx-c39fa1f8 job-1042/job-138603 ctx-c17ce6fc) copyAsync inspecting src type TEMPLATE copyAsync inspecting dest type VOLUME 2017-06-20 12:15:48,386 DEBUG [c.c.a.t.Request] (Work-Job-Executor-82:ctx-c39fa1f8 job-1042/job-138603 ctx-c17ce6fc) Seq 16-3622864425242874354: Sending { Cmd , MgmtId: 345050411715, via: 16(Flex-Xen6.flexhost.local), Ver: v1, Flags: 100111, [{"org.apache.cloudstack.storage.command.CopyCommand":{"srcTO":{"org.apache.cloudstack.storage.to.TemplateObjectTO":{"path":"ab6f3bcd-4c3e-4a7a-9f8b-45a822dbaaaf","origUrl":"http://download.cloud.com/templates/4.5/systemvm64template-4.5-xen.vhd.bz2","uuid":"8a4039f2-bb71-11e4-8c76-0050569b1662","id":1,"format":"VHD","accountId":1,"checksum":"2b15ab4401c2d655264732d3fc600241","hvm":false,"displayText":"SystemVM Template (XenServer)","imageDataStore":{"org.apache.cloudstack.storage.to.PrimaryDataStoreTO":{"uuid":"FlexSAN2-LUN0","id":5,"poolType":"PreSetup","host":"localhost","path":"/FlexSAN2-LUN0","port":0,"url":"PreSetup://localhost/FlexSAN2-LUN0/?ROLE=Primary&STOREUUID=FlexSAN2-LUN0"}},"name":"routing-1","hypervisorType":"XenServer"}},"destTO":{"org.apache.cloudstack.storage.to.VolumeObjectTO":{"uuid":"4dba9def-2657-430e-8cd8-9369aebcaa25","volumeType":"ROOT","dataStore":{"org.apache.cloudstack.storage.to.PrimaryDataStoreTO":{"uuid":"FlexSAN2-LUN0","id":5,"poolType":"PreSetup","host":"localhost","path":"/FlexSAN2-LUN0","port":0,"url":"PreSetup://localhost/FlexSAN2-LUN0/?ROLE=Primary&STOREUUID=FlexSAN2-LUN0"}},"name":"ROOT-20685","size":2689602048,"volumeId":25604,"vmName":"s-20685-VM","accountId":1,"format":"VHD","provisioningType":"THIN","id":25604,"deviceId":0,"hypervisorType":"XenServer"}},"executeInSequence":true,"options":{},"wait":0}}] } 2017-06-20 12:15:48,386 DEBUG [c.c.a.t.Request] (Work-Job-Executor-82:ctx-c39fa1f8 job-1042/job-138603 ctx-c17ce6fc) Seq 16-3622864425242874354: Executing: { Cmd , MgmtId: 345050411715, via: 16(Flex-Xen6.flexhost.local), Ver: v1, Flags: 100111, [{"org.apache.cloudstack.storage.command.CopyCommand":{"srcTO":{"org.apache.cloudstack.storage.to.TemplateObjectTO":{"path":"ab6f3bcd-4c3e-4a7a-9f8b-45a822dbaaaf","origUrl":"http://download.cloud.com/templates/4.5/systemvm64template-4.5-xen.vhd.bz2","uuid":"8a4039f2-bb71-11e4-8c76-0050569b1662","id":1,"format":"VHD","accountId":1,"checksum":"2b15ab4401c2d655264732d3fc600241","hvm":false,"displayText":"SystemVM Template (XenServer)","imageDataStore":{"org.apache.cloudstack.storage.to.PrimaryDataStoreTO":{"uuid":"FlexSAN2-LUN0","id":5,"poolType":"PreSetup","host":"localhost","path":"/FlexSAN2-LUN0","port":0,"url":"PreSetup://localhost/FlexSAN2-LUN0/?ROLE=Primary&STOREUUID=FlexSAN2-LUN0"}},"name":"routing-1","hypervisorType":"XenServer"}},"destTO":{"org.apache.cloudstack.storage.to.VolumeObjectTO":{"uuid":"4dba9def-2657-430e-8cd8-9369aebcaa25","volumeType":"ROOT","dataStore":{"org.apache.cloudstack.storage.to.PrimaryDataStoreTO":{"uuid":"FlexSAN2-LUN0","id":5,"poolType":"PreSetup","host":"localhost","path":"/FlexSAN2-LUN0","port":0,"url":"PreSetup://localhost/FlexSAN2-LUN0/?ROLE=Primary&STOREUUID=FlexSAN2-LUN0"}},"name":"ROOT-20685","size":2689602048,"volumeId":25604,"vmName":"s-20685-VM","accountId":1,"format":"VHD","provisioningType":"THIN","id":25604,"deviceId":0,"hypervisorType":"XenServer"}},"executeInSequence":true,"options":{},"wait":0}}] } 2017-06-20 12:15:48,386 DEBUG [c.c.a.m.DirectAgentAttache] (DirectAgent-74:ctx-0acdd419) Seq 16-3622864425242874354: Executing request 2017-06-20 12:15:48,387 DEBUG [c.c.n.g.PodBasedNetworkGuru] (Work-Job-Executor-88:ctx-c51dafa0 job-342/job-138604 ctx-75edebb0) Allocated a nic NicProfile[49817-12662-629b85e7-ce19-4568-9df7-143c76d24300-10.90.2.204-null for VM[ConsoleProxy|v-12662-VM] So how do I check UUID's to validate that they are correct ? 2017-06-20 12:15:48,391 DEBUG [c.c.h.x.r.XenServerStorageProcessor] (DirectAgent-74:ctx-0acdd419) Catch Exception com.xensource.xenapi.Types$UuidInvalid :VDI getByUuid for uuid: ab6f3bcd-4c3e-4a7a-9f8b-45a822dbaaaf failed due to The uuid you supplied was invalid. 2017-06-20 12:15:48,391 WARN [c.c.h.x.r.XenServerStorageProcessor] (DirectAgent-74:ctx-0acdd419) Unable to create volume; Pool=volumeTO[uuid=4dba9def-2657-430e-8cd8-9369aebcaa25|path=null|datastore=PrimaryDataStoreTO[uuid=FlexSAN2-LUN0|name=null|id=5|pooltype=PreSetup]]; Disk: com.cloud.utils.exception.CloudRuntimeException: Catch Exception com.xensource.xenapi.Types$UuidInvalid :VDI getByUuid for uuid: ab6f3bcd-4c3e-4a7a-9f8b-45a822dbaaaf failed due to The uuid you supplied was invalid. Jeremy -----Original Message----- From: Jeremy Peterson [mailto:jpeter...@acentek.net] Sent: Thursday, June 15, 2017 4:20 PM To: users@cloudstack.apache.org; S. Brüseke - proIO GmbH <s.brues...@proio.com> Subject: RE: Recreating SystemVM's What type of networking are you using on the XenServers? XenServers are connected with 6 nic's per host connected to separate nexus 5k switches NIC 0 and NIC 1 are Bond 0+1 10Gb nics NIC 2 and NIC 3 are Bond 2+3 10Gb nics NIC 4 and NIC 5 are Bond 4+5 2Gb nics Cloudstack is running Advanced networking Bond 0+1 is primary storage Bond 2+3 is secondary storage Bond 4+5 is Management What version of os does the ms run on? CentOS release 6.9 (Final) What are the systemvm templates defined in your env? http://cloudstack.apt-get.eu/systemvm/4.5/systemvm64template-4.5-xen.vhd.bz2 What is the version of the systemvm.iso? Successfully installed system VM template to /secondary/template/tmpl/1/1/ I just reinstalled systemvm's from the above 4.5-xen.vhd What is the capacity you have in your (test) environment? This is a production enviroment and currently cloudstack shows the following. Public IP Addresses 61% VLAN 35% Management IP Addresses 20% Primary Storage 44% CPU 21% Memory 5% Of cource Secondary Storage shows 0% What is the host os version for the hypervisors? XenServer 6.5 SP1 What is the management network range? management.network.cidr 10.90.1.0/24 What are the other physical networks? ?? Not sure what more you need What storage do you use? Primary - ISCSI Secondary - NFS Is it reachable from the systemvm? All of my CS management servers have internet access Is the big bad internet reachable for your SSVM’s public interface? My SSVM does not go online but yes the public network is the same as the VR public vlan and all instances behind VR's are connected to the internet at this time Jeremy -----Original Message----- From: Daan Hoogland [mailto:daan.hoogl...@shapeblue.com] Sent: Thursday, June 15, 2017 9:34 AM To: users@cloudstack.apache.org; S. Brüseke - proIO GmbH <s.brues...@proio.com> Subject: Re: Recreating SystemVM's Your problem might be like what Swen says, Jeremy but also a wrong systemvm offering or a fault in your management network definition. I am going to sum up some trivialities so bear with me; What type of networking are you using on the XenServers? What version of os does the ms run on? What are the systemvm templates defined in your env? What is the version of the systemvm.iso? What is the capacity you have in your (test) environment? What is the host os version for the hypervisors? What is the management network range? What are the other physical networks? What storage do you use? Is it reachable from the systemvm? Is the big bad internet reachable for your SSVM’s public interface? And of course, How is the weather, where you are at? I am not sure any of these question is going to lead you in the right direction but one of them should. On 15/06/17 13:56, "S. Brüseke - proIO GmbH" <s.brues...@proio.com> wrote: I once did have some similar problem with my systemvms and my root cause was that in the global settings it referred to the wrong systemvm template. I am not sure if this helps you, but wanted to tell you. Mit freundlichen Grüßen / With kind regards, Swen -----Ursprüngliche Nachricht----- Von: Jeremy Peterson [mailto:jpeter...@acentek.net] Gesendet: Donnerstag, 15. Juni 2017 01:55 An: users@cloudstack.apache.org Betreff: RE: Recreating SystemVM's Hahaha. The best response ever. I dug through these emails and someone had soft of the same log messages cannot attach network and blamed xenserver. Ok I'm cool with that but why oh why is it only system vms? Jeremy ________________________________________ From: Imran Ahmed [im...@eaxiom.net] Sent: Wednesday, June 14, 2017 6:22 PM To: users@cloudstack.apache.org Subject: RE: Recreating SystemVM's Yes, -----Original Message----- From: Jeremy Peterson [mailto:jpeter...@acentek.net] Sent: Wednesday, June 14, 2017 9:59 PM To: users@cloudstack.apache.org Subject: RE: Recreating SystemVM's Is there anyone out there reading these messages? Am I just not seeing responses? Jeremy -----Original Message----- From: Jeremy Peterson [mailto:jpeter...@acentek.net] Sent: Wednesday, June 14, 2017 8:12 AM To: users@cloudstack.apache.org Subject: RE: Recreating SystemVM's I opened an issue since this is still an issue. CLOUDSTACK-9960 Jeremy -----Original Message----- From: Jeremy Peterson [mailto:jpeter...@acentek.net] Sent: Sunday, June 11, 2017 9:10 AM To: users@cloudstack.apache.org Subject: Re: Recreating SystemVM's Any other suggestions? I am going to be scheduling to run XenServer updates. But this all points back to CANNOT_ATTACH_NETWORk. I've verified nothing is active on the Public IP space that those two VM's were living on. Jeremy ________________________________________ From: Jeremy Peterson <jpeter...@acentek.net> Sent: Friday, June 9, 2017 9:58 AM To: users@cloudstack.apache.org Subject: RE: Recreating SystemVM's I see the vm's try to create on a host that I just removed from maintenance mode to install updates and here are the logs I don't see anything that sticks out to me as a failure message. Jun 9 09:53:54 Xen3 SM: [13068] ['ip', 'route', 'del', '169.254.0.0/16'] Jun 9 09:53:54 Xen3 SM: [13068] pread SUCCESS Jun 9 09:53:54 Xen3 SM: [13068] ['ifconfig', 'xapi12', '169.254.0.1', 'netmask', '255.255.0.0'] Jun 9 09:53:54 Xen3 SM: [13068] pread SUCCESS Jun 9 09:53:54 Xen3 SM: [13068] ['ip', 'route', 'add', '169.254.0.0/16', 'dev', 'xapi12', 'src', '169.254.0.1'] Jun 9 09:53:54 Xen3 SM: [13068] pread SUCCESS Jun 9 09:53:54 Xen3 SM: [13071] ['ip', 'route', 'del', '169.254.0.0/16'] Jun 9 09:53:54 Xen3 SM: [13071] pread SUCCESS Jun 9 09:53:54 Xen3 SM: [13071] ['ifconfig', 'xapi12', '169.254.0.1', 'netmask', '255.255.0.0'] Jun 9 09:53:54 Xen3 SM: [13071] pread SUCCESS Jun 9 09:53:54 Xen3 SM: [13071] ['ip', 'route', 'add', '169.254.0.0/16', 'dev', 'xapi12', 'src', '169.254.0.1'] Jun 9 09:53:54 Xen3 SM: [13071] pread SUCCESS Jun 9 09:54:00 Xen3 SM: [13115] on-slave.multi: {'vgName': 'VG_XenStorage-469b6dcd-8466-3d03-de0e-cc3983e1b6e2', 'lvName1': 'VHD-633338a7-6c40-4aa6-b88e-c798b6fdc04d', 'action1': 'deactivateNoRefcount', 'action2': 'cleanupLock', 'uuid2': '633338a7-6c40-4aa6-b88e-c798b6fdc04d', 'ns2': 'lvm-469b6dcd-8466-3d03-de0e-cc3983e1b6e2'} Jun 9 09:54:00 Xen3 SM: [13115] LVMCache created for VG_XenStorage-469b6dcd-8466-3d03-de0e-cc3983e1b6e2 Jun 9 09:54:00 Xen3 SM: [13115] on-slave.action 1: deactivateNoRefcount Jun 9 09:54:00 Xen3 SM: [13115] LVMCache: will initialize now Jun 9 09:54:00 Xen3 SM: [13115] LVMCache: refreshing Jun 9 09:54:00 Xen3 SM: [13115] ['/usr/sbin/lvs', '--noheadings', '--units', 'b', '-o', '+lv_tags', '/dev/VG_XenStorage-469b6dcd-8466-3d03-de0e-cc3983e1b6e2'] Jun 9 09:54:00 Xen3 SM: [13115] pread SUCCESS Jun 9 09:54:00 Xen3 SM: [13115] ['/usr/sbin/lvchange', '-an', '/dev/VG_XenStorage-469b6dcd-8466-3d03-de0e-cc3983e1b6e2/VHD-633338a7-6c40-4 aa6-b88e-c798b6fdc04d'] Jun 9 09:54:00 Xen3 SM: [13115] pread SUCCESS Jun 9 09:54:00 Xen3 SM: [13115] ['/sbin/dmsetup', 'status', 'VG_XenStorage--469b6dcd--8466--3d03--de0e--cc3983e1b6e2-VHD--633338a7--6c40 --4aa6--b88e--c798b6fdc04d'] Jun 9 09:54:00 Xen3 SM: [13115] pread SUCCESS Jun 9 09:54:00 Xen3 SM: [13115] on-slave.action 2: cleanupLock Jun 9 09:54:16 Xen3 SM: [13230] ['ip', 'route', 'del', '169.254.0.0/16'] Jun 9 09:54:16 Xen3 SM: [13230] pread SUCCESS Jun 9 09:54:16 Xen3 SM: [13230] ['ifconfig', 'xapi12', '169.254.0.1', 'netmask', '255.255.0.0'] Jun 9 09:54:16 Xen3 SM: [13230] pread SUCCESS Jun 9 09:54:16 Xen3 SM: [13230] ['ip', 'route', 'add', '169.254.0.0/16', 'dev', 'xapi12', 'src', '169.254.0.1'] Jun 9 09:54:16 Xen3 SM: [13230] pread SUCCESS Jun 9 09:54:19 Xen3 updatempppathd: [15446] The garbage collection routine returned: 0 Jun 9 09:54:23 Xen3 SM: [13277] ['ip', 'route', 'del', '169.254.0.0/16'] Jun 9 09:54:23 Xen3 SM: [13277] pread SUCCESS Jun 9 09:54:23 Xen3 SM: [13277] ['ifconfig', 'xapi12', '169.254.0.1', 'netmask', '255.255.0.0'] Jun 9 09:54:23 Xen3 SM: [13277] pread SUCCESS Jun 9 09:54:23 Xen3 SM: [13277] ['ip', 'route', 'add', '169.254.0.0/16', 'dev', 'xapi12', 'src', '169.254.0.1'] Jun 9 09:54:23 Xen3 SM: [13277] pread SUCCESS Jeremy -----Original Message----- From: Jeremy Peterson [mailto:jpeter...@acentek.net] Sent: Friday, June 9, 2017 9:53 AM To: users@cloudstack.apache.org Subject: RE: Recreating SystemVM's I am checking SMlog now on all hosts. Jeremy -----Original Message----- From: Rajani Karuturi [mailto:raj...@apache.org] Sent: Friday, June 9, 2017 9:00 AM To: Users <users@cloudstack.apache.org> Subject: Re: Recreating SystemVM's on xenserver log, did you check what is causing " HOST_CANNOT_ATTACH_NETWORK"? ~Rajani http://cloudplatform.accelerite.com/ On Fri, Jun 9, 2017 at 7:00 PM, Jeremy Peterson <jpeter...@acentek.net> wrote: > 08:28:43 select * from vm_instance where name like 's-%' limit > 10000 7481 row(s) returned 0.000 sec / 0.032 sec > > All vm's 'state' returned Destoryed outside of the current vm 7873 > which is in a Stopped state but that goes Destroyed and a new get created. > > Any other suggestions? > > Jeremy > > > -----Original Message----- > From: Jeremy Peterson [mailto:jpeter...@acentek.net] > Sent: Thursday, June 8, 2017 12:47 AM > To: users@cloudstack.apache.org > Subject: Re: Recreating SystemVM's > > I'll make that change in the am. > > Today I put a host in maintence and rebooted because proxy and > secstore vm were constantly being created on that host and still no change. > > Let you know tomorrow. > > Jeremy > > > Sent from my Verizon, Samsung Galaxy smartphone > > > -------- Original message -------- > From: Rajani Karuturi <raj...@apache.org> > Date: 6/8/17 12:07 AM (GMT-06:00) > To: Users <users@cloudstack.apache.org> > Subject: Re: Recreating SystemVM's > > Did you check SMLog on xenserver? > unable to destroy task(com.xensource.xenapi.Task@256829a8) on > host(b34f086e-fabf-471e-9feb-8f54362d7d0f) due to You gave an invalid > object reference. The object may have recently been deleted. The > class parameter gives the type of reference given, and the handle > parameter echoes the bad value given. > > Looks like Destroy of SSVM failed. What state is SSVM in? mark it as > Destroyed in cloud DB and wait for cloudstack to create a new SSVM. > > ~Rajani > http://cloudplatform.accelerite.com/ > > On Thu, Jun 8, 2017 at 1:11 AM, Jeremy Peterson > <jpeter...@acentek.net> > wrote: > > > Probably agreed. > > > > But I ran toolstack restart on all hypervisors and v-3193 just tried > > to create and fail along with s-5398. > > > > The PIF error went away. But VM's are still recreating > > > > https://pastebin.com/4n4xBgMT > > > > New log from this afternoon. > > > > My catalina.out is over 4GB > > > > Jeremy > > > > > > -----Original Message----- > > From: Makrand [mailto:makrandsa...@gmail.com] > > Sent: Wednesday, June 7, 2017 12:52 AM > > To: users@cloudstack.apache.org > > Subject: Re: Recreating SystemVM's > > > > Hi there, > > > > Looks more like hypervisor issue. > > > > Just run *xe-toolstack-restart* on hosts where these VMs are trying > > to start or if you don't have too many hosts, better run on all > > members including master. most of i/o related issues squared off by > > toolstack bounce. > > > > -- > > Makrand > > > > > > On Wed, Jun 7, 2017 at 3:01 AM, Jeremy Peterson > > <jpeter...@acentek.net> > > wrote: > > > > > Ok so I pulled this from Sunday morning. > > > > > > https://pastebin.com/nCETw1sC > > > > > > > > > errorInfo: [HOST_CANNOT_ATTACH_NETWORK, > > > OpaqueRef:65d0c844-bd70-81e9-4518-8809e1dc0ee7, > > > OpaqueRef:0093ac3f-9f3a-37e1-9cdb-581398d27ba2] > > > > > > XenServer error. > > > > > > Now this still gets me because all of the other VM's launched just > fine. > > > > > > Going into XenCenter I see an error at the bottom This PIF is a > > > bond slave and cannot be plugged. > > > > > > ??? > > > > > > If I go to networking on the hosts I see the storage vlans and > > > bonds are all there. > > > > > > I see my GUEST-PUB bond is there and LACP is setup correct. > > > > > > Any suggestions ? > > > > > > > > > Jeremy > > > > > > > > > -----Original Message----- > > > From: Jeremy Peterson [mailto:jpeter...@acentek.net] > > > Sent: Tuesday, June 6, 2017 9:23 AM > > > To: users@cloudstack.apache.org > > > Subject: RE: Recreating SystemVM's > > > > > > Thank you all for those responses. > > > > > > I'll comb through my management-server.log and post a pastebin if > > > I'm scratching my head. > > > > > > Jeremy > > > > > > -----Original Message----- > > > From: Rajani Karuturi [mailto:raj...@apache.org] > > > Sent: Tuesday, June 6, 2017 6:53 AM > > > To: users@cloudstack.apache.org > > > Subject: Re: Recreating SystemVM's > > > > > > If the zone is enabled, cloudstack should recreate them automatically. > > > > > > ~ Rajani > > > > > > http://cloudplatform.accelerite.com/ > > > > > > On June 6, 2017 at 11:37 AM, Erik Weber (terbol...@gmail.com) > > > wrote: > > > > > > CloudStack should recreate automatically, check the mgmt server > > > logs for hints of why it doesn't happen. > > > > > > -- > > > Erik > > > > > > tir. 6. jun. 2017 kl. 04.29 skrev Jeremy Peterson > > > <jpeter...@acentek.net>: > > > > > > I had an issue Sunday morning with cloudstack 4.9.0 and xenserver > 6.5.0. > > > My hosts stop sending LACP PDU's and caused a network drop to > > > iSCSI primary storage. > > > > > > So all my instances recovered via HA enabled. > > > > > > But my console proxy and secondary storage system VM's got stuck > > > in a boot state that would not power on. > > > > > > At this time they are expunged and gone. > > > > > > How do I tell cloudstack-management to recreate system VM's? > > > > > > I'm drawing a blank since deploying CS two years ago and just > > > keeping things running and adding hosts and more storage > > > everything has been so stable. > > > > > > Jeremy > > > > > > - proIO GmbH - Geschäftsführer: Swen Brüseke Sitz der Gesellschaft: Frankfurt am Main USt-IdNr. DE 267 075 918 Registergericht: Frankfurt am Main - HRB 86239 Diese E-Mail enthält vertrauliche und/oder rechtlich geschützte Informationen. Wenn Sie nicht der richtige Adressat sind oder diese E-Mail irrtümlich erhalten haben, informieren Sie bitte sofort den Absender und vernichten Sie diese Mail. Das unerlaubte Kopieren sowie die unbefugte Weitergabe dieser Mail sind nicht gestattet. This e-mail may contain confidential and/or privileged information. If you are not the intended recipient (or have received this e-mail in error) please notify the sender immediately and destroy this e-mail. Any unauthorized copying, disclosure or distribution of the material in this e-mail is strictly forbidden. daan.hoogl...@shapeblue.com www.shapeblue.com 53 Chandos Place, Covent Garden, London WC2N 4HSUK @shapeblue