On 09/16/2013 07:16 PM, Marcus Sorensen wrote:
I agree, it makes things more complicated. However, updating libvirt should be a maintenance mode thing; I've seen it lose track of VMs between major libvirt upgrades and have to kill/restart all vms to reregister them with libvirt, best to be in maintenance. If I remember right, the issues with persistent storage definitiions were primarily related to failure scenarios, things like hard power off of host, host would build up dozens of primary storage definitions that it didn't need, creating unnecessary dependencies and potential unnecessary outages (see other email about what happens when storage goes down). If edison has a patch then that's great. One idea I head was that if the pool fails to register due to mountpoint already existing, we could use an alternate mount point to register it. Old stuff will keep working, new stuff will continue.
No! That will break migrations to machines which don't have that mountpoint. Backing devices with QCOW2 will also break on the longer run.
Please do not do that. Wido
On Mon, Sep 16, 2013 at 3:43 AM, Wei Zhou (JIRA) <[email protected]> wrote:[ https://issues.apache.org/jira/browse/CLOUDSTACK-3565?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13768197#comment-13768197 ] Wei Zhou commented on CLOUDSTACK-3565: -------------------------------------- Marcus, The change of the host to maintenance one by one makes upgrade/change complicated. Edison has changed the source to support re-creating storage pool in the fix of CLOUDSTACK-2729. However, the test failed in my environment. By the way, I did not see any other issue when I used cloudstack 4.0 while the storage pools are persistent. I do not know what issue will be leader by it.Restarting libvirtd service leading to destroy storage pool ----------------------------------------------------------- Key: CLOUDSTACK-3565 URL: https://issues.apache.org/jira/browse/CLOUDSTACK-3565 Project: CloudStack Issue Type: Bug Security Level: Public(Anyone can view this level - this is the default.) Components: KVM Affects Versions: 4.2.0 Environment: KVM Branch 4.2 Reporter: Rayees Namathponnan Assignee: Marcus Sorensen Priority: Blocker Labels: documentation Fix For: 4.2.0 Steps to reproduce Step 1 : Create cloudstack step in kvm Step 2 : From kvm host check "virsh pool-list" Step 3: Stop and start libvirtd service Step 4 : Check "virsh pool-list" Actual Result "virsh pool-list" is blank after restart libvird service [root@Rack2Host12 agent]# virsh pool-list Name State Autostart ----------------------------------------- 41b632b5-40b3-3024-a38b-ea259c72579f active no 469da865-0712-4d4b-a4cf-a2d68f99f1b6 active no fff90cb5-06dd-33b3-8815-d78c08ca01d9 active no [root@Rack2Host12 agent]# service cloudstack-agent stop Stopping Cloud Agent: [root@Rack2Host12 agent]# virsh pool-list Name State Autostart ----------------------------------------- 41b632b5-40b3-3024-a38b-ea259c72579f active no 469da865-0712-4d4b-a4cf-a2d68f99f1b6 active no fff90cb5-06dd-33b3-8815-d78c08ca01d9 active no [root@Rack2Host12 agent]# virsh list Id Name State ---------------------------------------------------- [root@Rack2Host12 agent]# service libvirtd stop Stopping libvirtd daemon: [ OK ] [root@Rack2Host12 agent]# service libvirtd start Starting libvirtd daemon: [ OK ] [root@Rack2Host12 agent]# virsh pool-list Name State Autostart ------------------------------------------- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
