[ 
https://issues.apache.org/jira/browse/CLOUDSTACK-8714?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Remi Bergsma resolved CLOUDSTACK-8714.
--------------------------------------
    Resolution: Fixed
      Assignee: Remi Bergsma

Fixed in 90feab18e028b8291169ad5171c999f9d8be3ec0

> Restore VM (Re-install VM) with enable.storage.migration set to false fails, 
> later fails to start up VM too
> -----------------------------------------------------------------------------------------------------------
>
>                 Key: CLOUDSTACK-8714
>                 URL: https://issues.apache.org/jira/browse/CLOUDSTACK-8714
>             Project: CloudStack
>          Issue Type: Bug
>      Security Level: Public(Anyone can view this level - this is the 
> default.) 
>    Affects Versions: 4.5.1
>            Reporter: Maneesha
>            Assignee: Remi Bergsma
>             Fix For: 4.6.0
>
>
> Environment: 
> ===========
> Advanced zone, 
> Hypervisor: XS, 
> Shared storage - multiple pools
> API: restoreVirtualMachine
> When we fire a Re-install VM, the allocator logic is kicking in for data 
> disks as well causing the data disks to possibly get migrated to a different 
> storage. If global config enable.storage.migration is set to false, the 
> migration would fail and Reset VM would also fail, although there's enough 
> space left in the existing storage pools to run the data disks.
> Later, when I try to start up the VM, (which has now gone into stopped 
> state), that also fails.
> Question is, why should we move around the data disks when we do Re-install 
> VM? Only the ROOT disk should be re-installed and re-deployed. The data disks 
> should remain as is. But the allocator logic is kicking in for all the disks 
> attached to the VM as well and in effect, may get migrated to different pools 
> in the cluster. We also add new entries in the DB for the data disks that got 
> migrated.
> If there are many number of data disks attached to the VM, we are spending a 
> lot of time just trying to un-necessarily move around the disks to different 
> pools. While we actually need only the ROOT disk to be re-installed.
> Finally, the VM also becomes un-recoverable since start VM fails.
> Steps:
> =====
> Have multiple pools in the cluster.
> Set enable.storage.migration = false
> Deploy a VM with data disk
> Re-install VM
> Watch for result. The data disks might get migrated to different storage if 
> the allocator decides to deploy them in different pool. It may take couple of 
> attempts to reproduce the issue, may not see it at the first time.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to