[ https://issues.apache.org/jira/browse/CLOUDSTACK-4208?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
venkata swamybabu budumuru closed CLOUDSTACK-4208. -------------------------------------------------- misunderstood the scenario. The behaviour I mentioned here is as expected hence closing the issue. > [ResetVM] ResetVM is not cleaning the old snapshots > ---------------------------------------------------- > > Key: CLOUDSTACK-4208 > URL: https://issues.apache.org/jira/browse/CLOUDSTACK-4208 > Project: CloudStack > Issue Type: Bug > Security Level: Public(Anyone can view this level - this is the > default.) > Components: Snapshot > Affects Versions: 4.2.0 > Environment: commit id # 7fced6460f13468369550c968670c7d0b9837949 > Reporter: venkata swamybabu budumuru > Assignee: Harikrishna Patnala > Priority: Critical > Fix For: 4.2.0 > > > Steps to reproduce: > 1. Have latest CloudStack setup with at least one advanced zone using a KVM > cluster of 2 hosts. > 2. As a non-ROOT admin, deploy a VM > 3. Take a snapshot > 4. Click on ResetVM option while the VM is running > Observations: > (i) I can see in the mgmt server logs for restoreVirtualMachine as mentioned > below. > 2013-08-09 15:18:37,991 DEBUG [cloud.api.ApiServlet] (catalina-exec-20:null) > ===START=== 10.252.192.25 -- GET > command=restoreVirtualMachine&virtualmachineid=5d7a81fe-65ef-4a17-8558-0df0ab7e02d8&response=json&sessionkey=BCQB7S0qwhF8ZaSpZ%2ByvCw%2F92d4%3D&_=1376041729404 > 2013-08-09 15:18:38,057 DEBUG [cloud.network.NetworkManagerImpl] > (Network-Scavenger-1:null) Lock is acquired for network Ntwk[236|Guest|8] as > a part of network shutdown > 2013-08-09 15:18:38,058 DEBUG [cloud.network.NetworkManagerImpl] > (Network-Scavenger-1:null) Network is not implemented: Ntwk[236|Guest|8] > 2013-08-09 15:18:38,061 DEBUG [cloud.network.NetworkManagerImpl] > (Network-Scavenger-1:null) Lock is released for network Ntwk[236|Guest|8] as > a part of network shutdown > 2013-08-09 15:18:38,108 DEBUG [cloud.async.AsyncJobManagerImpl] > (catalina-exec-20:null) submit async job-670 = [ > 2b6de978-2c71-4ad5-bc61-a0721d556f11 ], details: AsyncJobVO {id:670, userId: > 139, accountId: 151, sessionKey: null, instanceType: None, instanceId: null, > cmd: org.apache.cloudstack.api.command.user.vm.RestoreVMCmd, cmdOriginator: > null, cmdInfo: > {"response":"json","sessionkey":"BCQB7S0qwhF8ZaSpZ+yvCw/92d4\u003d","virtualmachineid":"5d7a81fe-65ef-4a17-8558-0df0ab7e02d8","cmdEventType":"VM.RESTORE","ctxUserId":"139","httpmethod":"GET","_":"1376041729404","ctxAccountId":"151","ctxStartEventId":"3210"}, > cmdVersion: 0, callbackType: 0, callbackAddress: null, status: 0, > processStatus: 0, resultCode: 0, result: null, initMsid: 7280707764394, > completeMsid: null, lastUpdated: null, lastPolled: null, created: null} > (ii) This action deleted the old volume with id 251 and created a new with > volume id 259 > mysql> select * from volumes where id=251\G > *************************** 1. row *************************** > id: 251 > account_id: 151 > domain_id: 3 > pool_id: 1 > last_pool_id: NULL > instance_id: NULL > device_id: NULL > name: ROOT-213 > uuid: c8951f7f-593a-4bb4-a9dc-2a9ed307c6f1 > size: 32212254720 > folder: NULL > path: 30a56120-caa3-424a-a291-3794ff74d5bf > pod_id: NULL > data_center_id: 1 > iscsi_name: NULL > host_ip: NULL > volume_type: ROOT > pool_type: NULL > disk_offering_id: 2 > template_id: 211 > first_snapshot_backup_uuid: NULL > recreatable: 0 > created: 2013-08-08 13:39:25 > attached: NULL > updated: 2013-08-09 09:48:49 > removed: 2013-08-09 09:48:49 > state: Expunged > chain_info: NULL > update_count: 8 > disk_type: NULL > vm_snapshot_chain_size: NULL > iso_id: 0 > display_volume: 0 > format: QCOW2 > min_iops: NULL > max_iops: NULL > 1 row in set (0.00 sec) > mysql> select * from volumes where instance_id=213\G > *************************** 1. row *************************** > id: 259 > account_id: 151 > domain_id: 3 > pool_id: 1 > last_pool_id: NULL > instance_id: 213 > device_id: 0 > name: ROOT-213 > uuid: b3331940-e778-4b36-8e74-ede4c8953c1b > size: 32212254720 > folder: NULL > path: 6c691045-c050-4321-b567-a6761b473ad8 > pod_id: NULL > data_center_id: 1 > iscsi_name: NULL > host_ip: NULL > volume_type: ROOT > pool_type: NULL > disk_offering_id: 2 > template_id: 211 > first_snapshot_backup_uuid: NULL > recreatable: 0 > created: 2013-08-09 09:48:49 > attached: 2013-08-09 09:48:49 > updated: 2013-08-09 09:48:50 > removed: NULL > state: Ready > chain_info: NULL > update_count: 2 > disk_type: NULL > vm_snapshot_chain_size: NULL > iso_id: 0 > display_volume: 0 > format: QCOW2 > min_iops: NULL > max_iops: NULL > 1 row in set (0.00 sec) > (iii) But, this event didn't trigger in deleting the old snapshots on old > volume i.e. 251 > mysql> select * from snapshot_store_ref\G > *************************** 1. row *************************** > id: 1 > store_id: 1 > snapshot_id: 1 > created: 2013-08-09 09:34:30 > last_updated: NULL > job_id: NULL > store_role: Primary > size: 0 > physical_size: 0 > parent_snapshot_id: 0 > install_path: > /mnt/5458182e-bfcb-351c-97ed-e7223bca2b8e/30a56120-caa3-424a-a291-3794ff74d5bf/51bb35f9-7b8c-4481-b45e-24baa8f8ce06 > state: Ready > update_count: 2 > ref_cnt: 0 > updated: 2013-08-09 09:34:32 > volume_id: 251 > *************************** 2. row *************************** > id: 2 > store_id: 1 > snapshot_id: 1 > created: 2013-08-09 09:34:32 > last_updated: NULL > job_id: NULL > store_role: Image > size: 0 > physical_size: 0 > parent_snapshot_id: 0 > install_path: snapshots/151/251/51bb35f9-7b8c-4481-b45e-24baa8f8ce06 > state: Ready > update_count: 2 > ref_cnt: 0 > updated: 2013-08-09 09:43:52 > volume_id: 251 > mysql> select * from snapshots\G > *************************** 1. row *************************** > id: 1 > data_center_id: 1 > account_id: 151 > domain_id: 3 > volume_id: 251 > disk_offering_id: 2 > status: BackedUp > path: NULL > name: 5d7a81fe-65ef-4a17-8558-0df0ab7e02d8_ROOT-213_20130809093429 > uuid: d3a403b9-d196-4c57-b2a8-85cf0c093833 > snapshot_type: 0 > type_description: MANUAL > size: 32212254720 > created: 2013-08-09 09:34:29 > removed: NULL > backup_snap_id: NULL > swift_id: NULL > sechost_id: NULL > prev_snap_id: NULL > hypervisor_type: KVM > version: 2.2 > s3_id: NULL > (iv) Checked the primary and I can see that the volume i.d 251 is deleted. > below is snippet from primary storage > [root@CentOS63Dev 1]# cd primary.campo.kvm.1.zone/ > [root@CentOS63Dev primary.campo.kvm.1.zone]# !ls > ls -lth | grep -i 30a > (v) Have checked on the secondary storage and I could still see the snapshot > existing thought the volume associated with is deleted. > [root@CentOS63Dev 251]# pwd > /tmp/1/secondary.campo.kvm.1/snapshots/151/251 > [root@CentOS63Dev 251]# ls -lth 51bb35f9-7b8c-4481-b45e-24baa8f8ce06 > -rw-r-----+ 1 root root 6.5G Aug 9 03:14 51bb35f9-7b8c-4481-b45e-24baa8f8ce06 > Here is global setting info > storage.cleanup.enabled = true > storage.cleanup.interval = 30 > Attaching all the logs along with db dump to the bug. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira