deepti dohare created CLOUDSTACK-356: ----------------------------------------
Summary: snapshot errors with multiple secondary storage in one zone Key: CLOUDSTACK-356 URL: https://issues.apache.org/jira/browse/CLOUDSTACK-356 Project: CloudStack Issue Type: Bug Components: Snapshot Affects Versions: pre-4.0.0 Reporter: deepti dohare Fix For: 4.1.0 BRIEF SYNOPSIS: ======================== Snapshots are failing inder some cases where the host has already mounted other secondary storage and snapshot command has been issued to it ENVIRONMENT ==================== CS 2.2.14 + XEN server DETAILED DESCRIPTION ==================== 1) Noticcing following error for some snapshots 2012-07-23 02:16:38,120 WARN [xen.resource.CitrixResourceBase] (DirectAgent-281:null) Task failed! Task record: uuid: f0642eda-6e40-8bca-1ffb-bbf52a0e307e status: FAILURE errorInfo: [XENAPI_PLUGIN_FAILURE, backupSnapshot, SROSError, Error reporting error, unknown key File /var/run/cloud_mount/1/snapshots/35/2521/a05404cd-e2c7-41cf-9972-aa019ab79494.vhd does not exist.] 2) Cluster has investigated the issue and they found that mount points are not identical for the xenservers to the secondary storages across a cluster (df) : * host 1 : 10.16.36.64:/RAID10-DATA01/secsatxen002/snapshots 2147483648 240044032 1907439616 12% /var/run/cloud_mount/1/snapshots * Host 2 : 10.16.36.64:/RAID10-DATA01/secsatxen001/snapshots 2147483648 569533440 1577950208 27% /var/run/cloud_mount/1/snapshots Therefore, the contents are not identical in the 2 secondary storages (this is right to my understanding, because multiple secondary storage permit to scale horizontally) : * Host 1 : * -rw-r--r--+ 1 nobody nobody 307M 2012-07-23 16:11 07ba46c7-0908-4402-b20a-5ddd29a65d17.vhd * -rw-r--r--+ 1 nobody nobody 301M 2012-07-18 02:18 1c378e6a-0eba-48d9-8545-ab9fa09b9ba0.vhd * -rw-r--r--+ 1 nobody nobody 59M 2012-07-19 02:18 3b7e8c19-ba9c-433e-ac18-f52d4ad42be9.vhd * -rw-r--r--+ 1 nobody nobody 97M 2012-07-20 02:19 4644362a-764d-4c66-a038-168127007a5f.vhd * -rw-r--r--+ 1 nobody nobody 47M 2012-07-23 02:18 72e54c85-8245-4b84-b4b5-5a1ef2b44e52.vhd * -rw-r--r--+ 1 nobody nobody 69M 2012-07-21 02:18 76301f19-fcc0-43ac-adab-cde2cf23bd87.vhd * -rw-r--r--+ 1 nobody nobody 57M 2012-07-22 02:18 a05404cd-e2c7-41cf-9972-aa019ab79494.vhd * Host 2 : empty A simple workaround is to migrate the corresponding instance to the host with the right mount. After further checking, this behavior is encountered on all clusters, only rarely because only applying to instances that : * Have a scheduled snapshot retention policy * Have either migrated, been restarted on a different host than previous run Though workaround is feasible but it is difficult to predict when the user seems this probelm. please check if there is any problem in cloudstack logic to ensure right mount secondary mount point is available and create it if it doesnt exist If this behaviour confirms to be a bug,please file one bug in jira. REPRO STEPS ================== You can try reproducing the scenario by havign two secondary storages and continuously taking snapshots EXPECTED BEHAVIOR ================== Cloudstack should create correct mount of secondary storage if it doesnt exist. It should never use other secondary storage mount point. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira