deepti dohare created CLOUDSTACK-356:
----------------------------------------
Summary: snapshot errors with multiple secondary storage in one
zone
Key: CLOUDSTACK-356
URL: https://issues.apache.org/jira/browse/CLOUDSTACK-356
Project: CloudStack
Issue Type: Bug
Components: Snapshot
Affects Versions: pre-4.0.0
Reporter: deepti dohare
Fix For: 4.1.0
BRIEF SYNOPSIS:
========================
Snapshots are failing inder some cases where the host has already mounted other
secondary storage and snapshot command has been issued to it
ENVIRONMENT
====================
CS 2.2.14 + XEN server
DETAILED DESCRIPTION
====================
1) Noticcing following error for some snapshots
2012-07-23 02:16:38,120 WARN [xen.resource.CitrixResourceBase]
(DirectAgent-281:null) Task failed! Task record: uuid:
f0642eda-6e40-8bca-1ffb-bbf52a0e307e
status: FAILURE
errorInfo: [XENAPI_PLUGIN_FAILURE, backupSnapshot, SROSError, Error reporting
error, unknown key File
/var/run/cloud_mount/1/snapshots/35/2521/a05404cd-e2c7-41cf-9972-aa019ab79494.vhd
does not exist.]
2) Cluster has investigated the issue and they found that mount points are not
identical for the xenservers to the secondary storages across a cluster (df) :
* host 1 : 10.16.36.64:/RAID10-DATA01/secsatxen002/snapshots 2147483648
240044032 1907439616 12% /var/run/cloud_mount/1/snapshots
* Host 2 : 10.16.36.64:/RAID10-DATA01/secsatxen001/snapshots 2147483648
569533440 1577950208 27% /var/run/cloud_mount/1/snapshots
Therefore, the contents are not identical in the 2 secondary storages (this is
right to my understanding, because multiple secondary storage permit to scale
horizontally) :
* Host 1 :
* -rw-r--r--+ 1 nobody nobody 307M 2012-07-23 16:11
07ba46c7-0908-4402-b20a-5ddd29a65d17.vhd
* -rw-r--r--+ 1 nobody nobody 301M 2012-07-18 02:18
1c378e6a-0eba-48d9-8545-ab9fa09b9ba0.vhd
* -rw-r--r--+ 1 nobody nobody 59M 2012-07-19 02:18
3b7e8c19-ba9c-433e-ac18-f52d4ad42be9.vhd
* -rw-r--r--+ 1 nobody nobody 97M 2012-07-20 02:19
4644362a-764d-4c66-a038-168127007a5f.vhd
* -rw-r--r--+ 1 nobody nobody 47M 2012-07-23 02:18
72e54c85-8245-4b84-b4b5-5a1ef2b44e52.vhd
* -rw-r--r--+ 1 nobody nobody 69M 2012-07-21 02:18
76301f19-fcc0-43ac-adab-cde2cf23bd87.vhd
* -rw-r--r--+ 1 nobody nobody 57M 2012-07-22 02:18
a05404cd-e2c7-41cf-9972-aa019ab79494.vhd
* Host 2 : empty
A simple workaround is to migrate the corresponding instance to the host with
the right mount. After further checking, this behavior is encountered on all
clusters, only rarely because only applying to instances that :
* Have a scheduled snapshot retention policy
* Have either migrated, been restarted on a different host than previous run
Though workaround is feasible but it is difficult to predict when the user
seems this probelm.
please check if there is any problem in cloudstack logic to ensure right mount
secondary mount point is available and create it if it doesnt exist
If this behaviour confirms to be a bug,please file one bug in jira.
REPRO STEPS
==================
You can try reproducing the scenario by havign two secondary storages and
continuously taking snapshots
EXPECTED BEHAVIOR
==================
Cloudstack should create correct mount of secondary storage if it doesnt exist.
It should never use other secondary storage mount point.
--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira