Here's the debugging I put in:

2013-08-07 17:39:12,286 DEBUG [kvm.storage.LibvirtStorageAdaptor]
(agentRequest-Handler-3:null) attempting to lock pool
2a9c418e-dd56-3f2c-9081-032b877d5644 with file
'/mnt/2a9c418e-dd56-3f2c-9081-032b877d5644/KVMFILELOCK/.lock'
2013-08-07 17:39:12,425 DEBUG [kvm.storage.LibvirtStorageAdaptor]
(agentRequest-Handler-3:null) attempting to lock pool
2fa10069-1fb3-4eea-8dc4-f773e7be0d19 with file
'/mnt/2fa10069-1fb3-4eea-8dc4-f773e7be0d19/KVMFILELOCK/.lock'
2013-08-07 17:39:12,479 DEBUG [cloud.agent.Agent]
(agentRequest-Handler-3:null) Seq 1-1276706825:  { Ans: , MgmtId:
90520735907663, via: 1, Ver: v1, Flags: 110,
[{"storage.PrimaryStorageDownloadAnswer":{"templateSize":0,"result":false,"details":"com.cloud.utils.exception.CloudRuntimeException:
Can not get file lock to refresh the pool
2fa10069-1fb3-4eea-8dc4-f773e7be0d19","wait":0}}] }
2013-08-07 17:39:12,536 DEBUG [cloud.agent.Agent]
(agentRequest-Handler-4:null) Request:Seq 1-1276706826:  { Cmd ,
MgmtId: 90520735907663, via: 1, Ver: v1, Flags: 100111,
[{"StopCommand":{"isProxy":false,"vmName":"s-57-VM","wait":0}}] }

Note that 2a9c418e-dd56-3f2c-9081-032b877d5644 is my secondary
storage, 2fa10069-1fb3-4eea-8dc4-f773e7be0d19 is local storage in
/var/lib/libvirt/images

On Wed, Aug 7, 2013 at 5:41 PM, Marcus Sorensen <shadow...@gmail.com> wrote:
> Here's the issue... local storage isn't mounted like remote storage is.
>
> This is invalid for local storage:
>
> String lockFile = _mountPoint + File.separator + pool.getUUIDString()
> + File.separator + _lockfile;
>
> Need to look at storage pool type and change lockfile to path for local 
> storage
>
> On Wed, Aug 7, 2013 at 5:18 PM, Marcus Sorensen (JIRA) <j...@apache.org> 
> wrote:
>>
>>     [ 
>> https://issues.apache.org/jira/browse/CLOUDSTACK-2729?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13732906#comment-13732906
>>  ]
>>
>> Marcus Sorensen commented on CLOUDSTACK-2729:
>> ---------------------------------------------
>>
>> Sorry, I think the KVMFILELOCK is left over on secondary storage
>> because it's unable to lock the primary storage, and doesn't clean up
>> after itself. I'm not sure why it can't take the lock out on primary
>> yet... troubleshooting that.
>>
>>
>>
>>> [Automation] Libvirt failed find primary  storage and VM deployment failed
>>> ----------------------------------------------------------------------------
>>>
>>>                 Key: CLOUDSTACK-2729
>>>                 URL: https://issues.apache.org/jira/browse/CLOUDSTACK-2729
>>>             Project: CloudStack
>>>          Issue Type: Bug
>>>      Security Level: Public(Anyone can view this level - this is the 
>>> default.)
>>>          Components: Storage Controller
>>>    Affects Versions: 4.2.0
>>>         Environment: Master branch build
>>> Automation environment - KVM
>>>            Reporter: Rayees Namathponnan
>>>            Assignee: edison su
>>>            Priority: Blocker
>>>             Fix For: 4.2.0
>>>
>>>         Attachments: CLOUDSTACK-2729.rar
>>>
>>>
>>> Failed to deploy VM in automation, environment created with , 2 hosts and 2 
>>> primary storages in a cluster
>>> 1) /export/home/rayees/SC_QA_AUTO4/primary2
>>> 2) /export/home/rayees/SC_QA_AUTO4/primary
>>> Libvirt failed to find the primary storage 
>>> /export/home/rayees/SC_QA_AUTO4/primary and VM deployment failed
>>> MS log
>>> 2013-05-28 21:11:44,540 DEBUG [agent.transport.Request] 
>>> (consoleproxy-1:null) Seq 4-936706756: Sending  { Cmd , MgmtId: 
>>> 29066118877352, via: 4, Ver: v1, Flags: 100111, 
>>> [{"StartCommand":{"vm":{"id":29,"name":"v-29-QA","type":"ConsoleProxy","cpus":1,"minSpeed":500,"maxSpeed":500,"minRam":1073741824,"maxRam":1073741824,"arch":"x86_64","os":"Debian
>>>  GNU/Linux 5.0 (32-bit)","bootArgs":" template=domP type=consoleproxy 
>>> host=10.223.49.195 port=8250 name=v-29-QA premium=true zone=1 pod=1 
>>> guid=Proxy.29 proxy_vm=29 disable_rp_filter=true eth2ip=10.223.122.73 
>>> eth2mask=255.255.255.192 gateway=10.223.122.65 eth0ip=169.254.0.154 
>>> eth0mask=255.255.0.0 eth1ip=10.223.50.96 eth1mask=255.255.255.192 
>>> mgmtcidr=10.223.49.192/26 localgw=10.223.50.65 internaldns1=10.223.110.254 
>>> dns1=72.52.126.11","rebootOnCrash":false,"enableHA":false,"limitCpuUse":false,"vncPassword":"d099568827911cef","params":{},"uuid":"5a146833-6a8c-44e5-83c0-50f34accf513","disks":[{"id":32,"name":"ROOT-29","mountPoint":"/export/home/rayees/SC_QA_AUTO4/primary","path":"f6f8d865-e9c0-4188-8a33-6c6383ca5075","size":276406784,"type":"ROOT","storagePoolType":"NetworkFilesystem","storagePoolUuid":"fff90cb5-06dd-33b3-8815-d78c08ca01d9","deviceId":0}],"nics":[{"deviceId":2,"networkRateMbps":-1,"defaultNic":true,"uuid":"21175978-96bd-4160-8228-8ad15aa40c66","ip":"10.223.122.73","netmask":"255.255.255.192","gateway":"10.223.122.65","mac":"06:2e:5a:00:00:42","dns1":"72.52.126.11","broadcastType":"Vlan","type":"Public","broadcastUri":"vlan://1221","isolationUri":"vlan://1221","isSecurityGroupEnabled":false},{"deviceId":0,"networkRateMbps":-1,"defaultNic":false,"uuid":"e0651452-a76e-4564-96b2-3d5d51e9bcd6","ip":"169.254.0.154","netmask":"255.255.0.0","gateway":"169.254.0.1","mac":"0e:00:a9:fe:00:9a","broadcastType":"LinkLocal","type":"Control","isSecurityGroupEnabled":false},{"deviceId":1,"networkRateMbps":-1,"defaultNic":false,"uuid":"0765c229-468e-4dfd-8382-24ac49791a8d","ip":"10.223.50.96","netmask":"255.255.255.192","gateway":"10.223.50.65","mac":"06:a8:e8:00:00:1d","broadcastType":"Native","type":"Management","isSecurityGroupEnabled":false}]},"hostIp":"10.223.50.66","wait":0}},{"check.CheckSshCommand":{"ip":"169.254.0.154","port":3922,"interval":6,"retries":100,"name":"v-29-QA","wait":0}}]
>>>  }
>>> 2013-05-28 21:11:44,552 DEBUG [agent.transport.Request] 
>>> (AgentManager-Handler-1:null) Seq 4-936706756: Processing:  { Ans: , 
>>> MgmtId: 29066118877352, via: 4, Ver: v1, Flags: 110, 
>>> [{"Answer":{"result":false,"details":"com.cloud.utils.exception.CloudRuntimeException:
>>>  org.libvirt.LibvirtException: Storage pool not found: no pool with 
>>> matching uuid\n\tat 
>>> com.cloud.hypervisor.kvm.storage.LibvirtStorageAdaptor.getStoragePool(LibvirtStorageAdaptor.java:380)\n\tat
>>>  
>>> com.cloud.hypervisor.kvm.storage.KVMStoragePoolManager.getStoragePool(KVMStoragePoolManager.java:72)\n\tat
>>>  
>>> com.cloud.hypervisor.kvm.resource.LibvirtComputingResource.createVbd(LibvirtComputingResource.java:3399)\n\tat
>>>  
>>> com.cloud.hypervisor.kvm.resource.LibvirtComputingResource.execute(LibvirtComputingResource.java:3293)\n\tat
>>>  
>>> com.cloud.hypervisor.kvm.resource.LibvirtComputingResource.executeRequest(LibvirtComputingResource.java:1187)\n\tat
>>>  com.cloud.agent.Agent.processRequest(Agent.java:525)\n\tat 
>>> com.cloud.agent.Agent$AgentRequestHandler.doTask(Agent.java:852)\n\tat 
>>> com.cloud.utils.nio.Task.run(Task.java:83)\n\tat 
>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1110)\n\tat
>>>  
>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:603)\n\tat
>>>  
>>> java.lang.Thread.run(Thread.java:679)\n","wait":0}},{"Answer":{"result":false,"details":"Stopped
>>>  by previous failure","wait":0}}] }
>>> 2013-05-28 21:11:44,552 DEBUG [agent.manager.AgentAttache] 
>>> (AgentManager-Handler-1:null) Seq 4-936706756: No more commands found
>>> 2013-05-28 21:11:44,552 DEBUG [agent.transport.Request] 
>>> (consoleproxy-1:null) Seq 4-936706756: Received:  { Ans: , MgmtId: 
>>> 29066118877352, via: 4, Ver: v1, Flags: 110, { Answer, Answer } }
>>> 2013-05-28 21:11:44,559 ERROR [cloud.vm.VirtualMachineManagerImpl] 
>>> (consoleproxy-1:null) Failed to start instance VM[ConsoleProxy|v-29-QA]
>>> com.cloud.utils.exception.CloudRuntimeException: Unable to get answer that 
>>> is of class com.cloud.agent.api.StartAnswer
>>>         at com.cloud.agent.manager.Commands.getAnswer(Commands.java:80)
>>>         at 
>>> com.cloud.vm.VirtualMachineManagerImpl.advanceStart(VirtualMachineManagerImpl.java:873)
>>>         at 
>>> com.cloud.vm.VirtualMachineManagerImpl.start(VirtualMachineManagerImpl.java:550)
>>>         at 
>>> com.cloud.vm.VirtualMachineManagerImpl.start(VirtualMachineManagerImpl.java:543)
>>>         at 
>>> com.cloud.consoleproxy.ConsoleProxyManagerImpl.startProxy(ConsoleProxyManagerImpl.java:561)
>>>         at 
>>> com.cloud.consoleproxy.ConsoleProxyManagerImpl.allocCapacity(ConsoleProxyManagerImpl.java:923)
>>>         at 
>>> com.cloud.consoleproxy.ConsoleProxyManagerImpl.expandPool(ConsoleProxyManagerImpl.java:1667)
>>>         at 
>>> com.cloud.consoleproxy.ConsoleProxyManagerImpl.expandPool(ConsoleProxyManagerImpl.java:157)
>>>         at 
>>> com.cloud.vm.SystemVmLoadScanner.loadScan(SystemVmLoadScanner.java:111)
>>> Agent log
>>> 2013-05-29 01:06:07,911 DEBUG [cloud.agent.Agent] 
>>> (agentRequest-Handler-4:null) Request:Seq 4-936706739:  { Cmd , MgmtId: 
>>> 29066118877352, via: 4, Ver: v1, Flags: 100111, 
>>> [{"StartCommand":{"vm":{"id":29,"name":"v-29-QA","type":"ConsoleProxy","cpus":1,"minSpeed":500,"maxSpeed":500,"minRam":1073741824,"maxRam":1073741824,"arch":"x86_64","os":"Debian
>>>  GNU/Linux 5.0 (32-bit)","bootArgs":" template=domP type=consoleproxy 
>>> host=10.223.49.195 port=8250 name=v-29-QA premium=true zone=1 pod=1 
>>> guid=Proxy.29 proxy_vm=29 disable_rp_filter=true eth2ip=10.223.122.73 
>>> eth2mask=255.255.255.192 gateway=10.223.122.65 eth0ip=169.254.1.243 
>>> eth0mask=255.255.0.0 eth1ip=10.223.50.93 eth1mask=255.255.255.192 
>>> mgmtcidr=10.223.49.192/26 localgw=10.223.50.65 internaldns1=10.223.110.254 
>>> dns1=72.52.126.11","rebootOnCrash":false,"enableHA":false,"limitCpuUse":false,"vncPassword":"d099568827911cef","params":{},"uuid":"5a146833-6a8c-44e5-83c0-50f34accf513","disks":[{"id":32,"name":"ROOT-29","mountPoint":"/export/home/rayees/SC_QA_AUTO4/primary","path":"f6f8d865-e9c0-4188-8a33-6c6383ca5075","size":276406784,"type":"ROOT","storagePoolType":"NetworkFilesystem","storagePoolUuid":"fff90cb5-06dd-33b3-8815-d78c08ca01d9","deviceId":0}],"nics":[{"deviceId":2,"networkRateMbps":-1,"defaultNic":true,"uuid":"21175978-96bd-4160-8228-8ad15aa40c66","ip":"10.223.122.73","netmask":"255.255.255.192","gateway":"10.223.122.65","mac":"06:2e:5a:00:00:42","dns1":"72.52.126.11","broadcastType":"Vlan","type":"Public","broadcastUri":"vlan://1221","isolationUri":"vlan://1221","isSecurityGroupEnabled":false},{"deviceId":0,"networkRateMbps":-1,"defaultNic":false,"uuid":"e0651452-a76e-4564-96b2-3d5d51e9bcd6","ip":"169.254.1.243","netmask":"255.255.0.0","gateway":"169.254.0.1","mac":"0e:00:a9:fe:01:f3","broadcastType":"LinkLocal","type":"Control","isSecurityGroupEnabled":false},{"deviceId":1,"networkRateMbps":-1,"defaultNic":false,"uuid":"0765c229-468e-4dfd-8382-24ac49791a8d","ip":"10.223.50.93","netmask":"255.255.255.192","gateway":"10.223.50.65","mac":"06:49:6a:00:00:1a","broadcastType":"Native","type":"Management","isSecurityGroupEnabled":false}]},"hostIp":"10.223.50.66","wait":0}},{"check.CheckSshCommand":{"ip":"169.254.1.243","port":3922,"interval":6,"retries":100,"name":"v-29-QA","wait":0}}]
>>>  }
>>> 2013-05-29 01:06:07,912 DEBUG [cloud.agent.Agent] 
>>> (agentRequest-Handler-4:null) Processing command: 
>>> com.cloud.agent.api.StartCommand
>>> 2013-05-29 01:06:07,923 WARN  [cloud.agent.Agent] 
>>> (agentRequest-Handler-4:null) Caught:
>>> com.cloud.utils.exception.CloudRuntimeException: 
>>> org.libvirt.LibvirtException: Storage pool not found: no pool with matching 
>>> uuid
>>>         at 
>>> com.cloud.hypervisor.kvm.storage.LibvirtStorageAdaptor.getStoragePool(LibvirtStorageAdaptor.java:380)
>>>         at 
>>> com.cloud.hypervisor.kvm.storage.KVMStoragePoolManager.getStoragePool(KVMStoragePoolManager.java:72)
>>>         at 
>>> com.cloud.hypervisor.kvm.resource.LibvirtComputingResource.createVbd(LibvirtComputingResource.java:3399)
>>>         at 
>>> com.cloud.hypervisor.kvm.resource.LibvirtComputingResource.execute(LibvirtComputingResource.java:3293)
>>>         at 
>>> com.cloud.hypervisor.kvm.resource.LibvirtComputingResource.executeRequest(LibvirtComputingResource.java:1187)
>>>         at com.cloud.agent.Agent.processRequest(Agent.java:525)
>>>         at com.cloud.agent.Agent$AgentRequestHandler.doTask(Agent.java:852)
>>>         at com.cloud.utils.nio.Task.run(Task.java:83)
>>>         at 
>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1110)
>>>         at 
>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:603)
>>> > create VM issuing create command with primary storage 
>>> > “/export/home/rayees/SC_QA_AUTO4/primary” but its not available in 
>>> > pool-list
>>> [root@Rack2Host11 ~]# virsh pool-list --all
>>> Name                 State      Autostart
>>> -----------------------------------------
>>> 1c9fcb9a-aece-453b-8510-4de4b4dbffdc active     no
>>> 41b632b5-40b3-3024-a38b-ea259c72579f active     no
>>> [root@Rack2Host11 ~]#
>>> [root@Rack2Host11 ~]#
>>> [root@Rack2Host11 ~]# virsh pool-dumpxml 
>>> 1c9fcb9a-aece-453b-8510-4de4b4dbffdc
>>> <pool type='dir'>
>>>   <name>1c9fcb9a-aece-453b-8510-4de4b4dbffdc</name>
>>>   <uuid>1c9fcb9a-aece-453b-8510-4de4b4dbffdc</uuid>
>>>   <capacity unit='bytes'>237777190912</capacity>
>>>   <allocation unit='bytes'>2098745344</allocation>
>>>   <available unit='bytes'>235678445568</available>
>>>   <source>
>>>   </source>
>>>   <target>
>>>     <path>/var/lib/libvirt/images</path>
>>>     <permissions>
>>>       <mode>0700</mode>
>>>       <owner>-1</owner>
>>>       <group>-1</group>
>>>     </permissions>
>>>   </target>
>>> </pool>
>>> [root@Rack2Host11 ~]# virsh pool-dumpxml 
>>> 41b632b5-40b3-3024-a38b-ea259c72579f
>>> <pool type='netfs'>
>>>   <name>41b632b5-40b3-3024-a38b-ea259c72579f</name>
>>>   <uuid>41b632b5-40b3-3024-a38b-ea259c72579f</uuid>
>>>   <capacity unit='bytes'>11810778316800</capacity>
>>>   <allocation unit='bytes'>9045745336320</allocation>
>>>   <available unit='bytes'>2765032980480</available>
>>>   <source>
>>>     <host name='10.223.110.232'/>
>>>     <dir path='/export/home/rayees/SC_QA_AUTO4/primary2'/>
>>>     <format type='auto'/>
>>>   </source>
>>>   <target>
>>>     <path>/mnt/41b632b5-40b3-3024-a38b-ea259c72579f</path>
>>>     <permissions>
>>>       <mode>0700</mode>
>>>       <owner>-1</owner>
>>>       <group>-1</group>
>>>     </permissions>
>>>   </target>
>>> </pool>
>>
>> --
>> This message is automatically generated by JIRA.
>> If you think it was sent incorrectly, please contact your JIRA administrators
>> For more information on JIRA, see: http://www.atlassian.com/software/jira

Reply via email to