[
https://issues.apache.org/jira/browse/CLOUDSTACK-3229?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13741863#comment-13741863
]
Thomas O'Dowd commented on CLOUDSTACK-3229:
-------------------------------------------
Ok. I blew away my devcloud instance and created it again from scratch. I also
blew away cloudstack and re-cloned it and checked out 4.2.
Still the same problem.
INFO [user.snapshot.CreateSnapshotCmd] (Job-Executor-2:job-9 = [
63c57cb9-5dea-4fb1-98cb-a17ec0fdd7d5 ]) VOLSS: createSnapshotCmd
starts:1376623660250
WARN [xen.resource.CitrixResourceBase] (DirectAgent-8:) callHostPlugin failed
for cmd: deleteSnapshotBackup with args secondaryStorageMountPath:
192.168.56.10:/opt/storage/cache/, localMountPoint:
/run/cloud_mount/49ab0a32-3113-3ee2-bb24-3e65acb96ce9, path: snapshots/2/3,
backupUUID: fdc22527-d4bc-4027-b3b7-018baa08b283, due to There was a failure
communicating with the plugin.
WARN [xen.resource.XenServerStorageProcessor] (DirectAgent-8:) BackupSnapshot
Failed due to callHostPlugin failed for cmd: deleteSnapshotBackup with args
secondaryStorageMountPath: 192.168.56.10:/opt/storage/cache/, localMountPoint:
/run/cloud_mount/49ab0a32-3113-3ee2-bb24-3e65acb96ce9, path: snapshots/2/3,
backupUUID: fdc22527-d4bc-4027-b3b7-018baa08b283, due to There was a failure
communicating with the plugin.
com.cloud.utils.exception.CloudRuntimeException: callHostPlugin failed for cmd:
deleteSnapshotBackup with args secondaryStorageMountPath:
192.168.56.10:/opt/storage/cache/, localMountPoint:
/run/cloud_mount/49ab0a32-3113-3ee2-bb24-3e65acb96ce9, path: snapshots/2/3,
backupUUID: fdc22527-d4bc-4027-b3b7-018baa08b283, due to There was a failure
communicating with the plugin.
at
com.cloud.hypervisor.xen.resource.CitrixResourceBase.callHostPlugin(CitrixResourceBase.java:4176)
at
com.cloud.hypervisor.xen.resource.XenServerStorageProcessor.deleteSnapshotBackup(XenServerStorageProcessor.java:1049)
at
com.cloud.hypervisor.xen.resource.XenServerStorageProcessor.backupSnapshot(XenServerStorageProcessor.java:1275)
at
com.cloud.storage.resource.StorageSubsystemCommandHandlerBase.execute(StorageSubsystemCommandHandlerBase.java:84)
at
com.cloud.storage.resource.StorageSubsystemCommandHandlerBase.handleStorageCommands(StorageSubsystemCommandHandlerBase.java:49)
at
com.cloud.hypervisor.xen.resource.CitrixResourceBase.executeRequest(CitrixResourceBase.java:617)
at
com.cloud.hypervisor.xen.resource.XcpOssResource.executeRequest(XcpOssResource.java:143)
at
com.cloud.agent.manager.DirectAgentAttache$Task.run(DirectAgentAttache.java:186)
at
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334)
at java.util.concurrent.FutureTask.run(FutureTask.java:166)
at
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:178)
at
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:292)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:724)
[13800] 2013-08-16 03:29:01.349974 #### VMOPS enter deleteSnapshotBackup
####
[13800] 2013-08-16 03:29:01.350068 Calling deleteSnapshotBackup with
{'localMountPoint': '/run/cloud_mount/49ab0a32-3113-3ee2-bb24-3e65acb96ce9',
'backupUUID': 'fdc22527-d4bc-4027-b3b7-018baa08b283',
'secondaryStorageMountPath': '192.168.56.10:/opt/storage/cache/', 'path':
'snapshots/2/3'}
[13807] 2013-08-16 03:29:01.488081 lock: acquired
/var/lock/sm/78f02bdd-21d3-b2f8-c05f-0c29b24b50ce/sr
[13807] 2013-08-16 03:29:01.488184 sr_detach {'sr_uuid':
'78f02bdd-21d3-b2f8-c05f-0c29b24b50ce', 'subtask_of':
'DummyRef:|07befe81-d4b4-47df-fa2f-cb309e275597|SR.detach', 'args': [],
'host_ref': 'OpaqueRef:0d19f3fa-7f04-543a-d906-aaf478fc574c', 'session_ref':
'OpaqueRef:d4ad234c-6bde-2808-b965-3ee8f419c75b', 'device_config': {'SRmaster':
'true', 'serverpath': '/opt/storage/cache/snapshots/2/3', 'server':
'192.168.56.10'}, 'command': 'sr_detach', 'sr_ref':
'OpaqueRef:9edf6bf2-4fa8-f996-f515-d46a0a4dbe88'}
[13807] 2013-08-16 03:29:01.488494 Aborting GC/coalesce
<13807> 2013-08-16 03:29:01.488546 === SR
78f02bdd-21d3-b2f8-c05f-0c29b24b50ce: abort ===
[13807] 2013-08-16 03:29:01.488649 lock: tried lock
/var/lock/sm/78f02bdd-21d3-b2f8-c05f-0c29b24b50ce/running, acquired: True
(exists: True)
<13807> 2013-08-16 03:29:01.488692 abort: releasing the process lock
[13807] 2013-08-16 03:29:01.488746 lock: released
/var/lock/sm/78f02bdd-21d3-b2f8-c05f-0c29b24b50ce/running
[13807] 2013-08-16 03:29:01.488806 ['umount',
'/var/run/sr-mount/78f02bdd-21d3-b2f8-c05f-0c29b24b50ce']
[13807] 2013-08-16 03:29:01.498655 SUCCESS
[13807] 2013-08-16 03:29:01.498846 lock: released
/var/lock/sm/78f02bdd-21d3-b2f8-c05f-0c29b24b50ce/sr
[13807] 2013-08-16 03:29:01.499139 lock: closed
/var/lock/sm/78f02bdd-21d3-b2f8-c05f-0c29b24b50ce/sr
[13807] 2013-08-16 03:29:01.502424 lock: closed
/var/lock/sm/78f02bdd-21d3-b2f8-c05f-0c29b24b50ce/running
[13846] 2013-08-16 03:29:15.631354 #### VMOPS enter gethostvmstats ####
[13846] 2013-08-16 03:29:15.635473 #### VMOPS exit gethostvmstats ####
[13853] 2013-08-16 03:29:18.378207 #### VMOPS enter gethostvmstats ####
[13853] 2013-08-16 03:29:18.382318 #### VMOPS exit gethostvmstats ####
I tried running df a few times when the snapshot was in progress and caught
this.
root@devcloud:~# df
Filesystem 1K-blocks Used
Available Use% Mounted on
rootfs 30277668 3685796
26591872 13% /
udev 10240 0
10240 0% /dev
tmpfs 148540 432
148108 1% /run
/dev/disk/by-uuid/c18fd62e-7d84-4d0e-b12e-a120b06d9e95 30277668 3685796
26591872 13% /
tmpfs 5120 0
5120 0% /run/lock
tmpfs 435920 0
435920 0% /run/shm
192.168.56.10:/opt/storage/cache/snapshots/2/3/ 30277888 3685888
26592000 13% /run/sr-mount/78f02bdd-21d3-b2f8-c05f-0c29b24b50ce
Is that the mount point you expect?
> Object_Store_Refactor - Snapshot fails due to an internal error
> ---------------------------------------------------------------
>
> Key: CLOUDSTACK-3229
> URL: https://issues.apache.org/jira/browse/CLOUDSTACK-3229
> Project: CloudStack
> Issue Type: Bug
> Security Level: Public(Anyone can view this level - this is the
> default.)
> Affects Versions: 4.2.0
> Environment: chrome on linux
> devcloud
> Cloudian or Amazon S3 Object store
> Reporter: Thomas O'Dowd
> Assignee: Min Chen
> Priority: Blocker
> Fix For: 4.2.0
>
> Attachments: SMlog, SMlog.last_5000_lines.txt
>
>
> Assuming initial devcloud state...
> I added a cache for the S3 storage like this.
> on devcloud machine as root:
> # mkdir /opt/storage/cache
> # vi /etc/exports (and append this line)
> /opt/storage/cache *(rw,no_subtree_check,no_root_squash,fsid=9999)
> # exportfs -a
> On Mgmt server GUI:
> 1. navigate to infrastructure -> secondary storage
> 2. delete the NFS SS.
> 3. add S3 storage for Cloudian (I used 60000 as the timeouts - assuming
> millis). I used the /opt/storage/cache thing as the s3 cache.
> 4. nav to templates
> 5. register a new template (I uploaded tinyLinux again as "mytiny" (5.3
> 64bit)).
> 6. confirm with s3cmd that 2 objects are now on S3.
> --------- s3 objects -------
> template/tmpl/1/1/routing-1/acton-systemvm-02062012.vhd.bz2
> 2013-06-27T03:01:46.203Z None 140616708 "b533e7b65219439ee7fca0146ddd7ffa-27"
> template/tmpl/2/201/201-2-ae9e9409-4c8e-3ad8-a62f-abec7a35fe26/tinylinux.vhd
> 2013-06-27T03:04:06.730Z None 50430464 "4afac316e865adf74ca1a8039fae7399-10"
> --------- s3 objects -------
> 7. I restarted the management server at this point which actually resulted in
> another object on S3.
> --------- the new s3 object -------
> template/tmpl/1/5/tiny Linux/ttylinux_pv.vhd 2013-06-27T03:43:26.494Z None
> 50430464 "4afac316e865adf74ca1a8039fae7399-10"
> --------- the new s3 object -------
> 8. Go to instance and create a new choosing the "mytiny" template which we
> registered.
> 9. launch it after selecting all defaults.
> 10. wait until it starts.
> 11. nav to storage. I see ROOT-8. Click on this to open.
> 12. click the camera to take the snapshot.
> after a pause I get a popup
> "Failed to create snapshot due to an internal error creating snapshot
> for volume 8"
> Also on the mgmt terminal I get the following log entry (only 1):
> INFO [user.snapshot.CreateSnapshotCmd] (Job-Executor-8:job-16) VOLSS:
> createSnapshotCmd starts:1372321251009
> If I check the "view snapshots" button under storage, I can however see the
> snapshot. It says its on primary. I'm expecting it to go to secondary storage
> though. Nothing is in the S3 logs and no snapshots.
> If I try to delete that snapshot from here I get this error in the logs:
> ERROR [cloud.async.AsyncJobManagerImpl] (Job-Executor-12:job-20) Unexpected
> exception while executing
> org.apache.cloudstack.api.command.user.snapshot.DeleteSnapshotCmd
> com.cloud.utils.exception.CloudRuntimeException: Failed to delete
> snapshot:com.cloud.exception.InvalidParameterValueException: Can't delete
> snapshotshot 4 due to it is not in BackedUp Status
> at
> com.cloud.storage.snapshot.SnapshotManagerImpl.deleteSnapshot(SnapshotManagerImpl.java:513)
> at
> com.cloud.utils.component.ComponentInstantiationPostProcessor$InterceptorDispatcher.intercept(ComponentInstantiationPostProcessor.java:125)
> at
> org.apache.cloudstack.api.command.user.snapshot.DeleteSnapshotCmd.execute(DeleteSnapshotCmd.java:96)
> at com.cloud.api.ApiDispatcher.dispatch(ApiDispatcher.java:155)
> at
> com.cloud.async.AsyncJobManagerImpl$1.run(AsyncJobManagerImpl.java:437)
> at
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
> at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334)
> at java.util.concurrent.FutureTask.run(FutureTask.java:166)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
> at java.lang.Thread.run(Thread.java:722)
> If I navigate to instance, my instance, and try to take a vm snapshot from
> here, I get a different pop-up which says:
> "There is other active volume snapshot tasks on the instance to which the
> volume is attached, please try again later"
> And I get an exception:
> ERROR [cloud.api.ApiServer] (352129314@qtp-2110413789-32:) unhandled
> exception executing api command: createVMSnapshot
> com.cloud.utils.exception.CloudRuntimeException: There is other active volume
> snapshot tasks on the instance to which the volume is attached, please try
> again later.
> at
> com.cloud.vm.snapshot.VMSnapshotManagerImpl.allocVMSnapshot(VMSnapshotManagerImpl.java:299)
> at
> org.apache.cloudstack.api.command.user.vmsnapshot.CreateVMSnapshotCmd.create(CreateVMSnapshotCmd.java:78)
> at
> com.cloud.api.ApiDispatcher.dispatchCreateCmd(ApiDispatcher.java:101)
> at com.cloud.api.ApiServer.queueCommand(ApiServer.java:475)
> at com.cloud.api.ApiServer.handleRequest(ApiServer.java:371)
> at com.cloud.api.ApiServlet.processRequest(ApiServlet.java:304)
> at com.cloud.api.ApiServlet.doGet(ApiServlet.java:66)
> at javax.servlet.http.HttpServlet.service(HttpServlet.java:707)
> at javax.servlet.http.HttpServlet.service(HttpServlet.java:820)
> at
> org.mortbay.jetty.servlet.ServletHolder.handle(ServletHolder.java:511)
> at
> org.mortbay.jetty.servlet.ServletHandler.handle(ServletHandler.java:401)
> at
> org.mortbay.jetty.security.SecurityHandler.handle(SecurityHandler.java:216)
> at
> org.mortbay.jetty.servlet.SessionHandler.handle(SessionHandler.java:182)
> at
> org.mortbay.jetty.handler.ContextHandler.handle(ContextHandler.java:766)
> at
> org.mortbay.jetty.webapp.WebAppContext.handle(WebAppContext.java:450)
> at
> org.mortbay.jetty.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:230)
> at
> org.mortbay.jetty.handler.HandlerCollection.handle(HandlerCollection.java:114)
> at
> org.mortbay.jetty.handler.HandlerWrapper.handle(HandlerWrapper.java:152)
> at org.mortbay.jetty.Server.handle(Server.java:326)
> at
> org.mortbay.jetty.HttpConnection.handleRequest(HttpConnection.java:542)
> at
> org.mortbay.jetty.HttpConnection$RequestHandler.headerComplete(HttpConnection.java:928)
> at org.mortbay.jetty.HttpParser.parseNext(HttpParser.java:549)
> at org.mortbay.jetty.HttpParser.parseAvailable(HttpParser.java:212)
> at org.mortbay.jetty.HttpConnection.handle(HttpConnection.java:404)
> at
> org.mortbay.io.nio.SelectChannelEndPoint.run(SelectChannelEndPoint.java:410)
> at
> org.mortbay.thread.QueuedThreadPool$PoolThread.run(QueuedThreadPool.java:582)
> There are no requests going to the S3 storage for the snap-shotting that I
> can see and its the only secondary storage that I have setup.
--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira