Re: [ovirt-users] Ovirt snapshot failing on one VM

2014-04-25 Thread Steve Dainard
Restarting vdsm and hosts didn't do anything helpful.

I was able to clone from latest snapshot, then live snapshot the new cloned
VM.

After upgrading engine to 3.4 and upgrading my hosts, I can now live
snapshot this VM.


*Steve *


On Thu, Apr 24, 2014 at 1:48 AM, Itamar Heim  wrote:

> On 04/23/2014 09:57 PM, R P Herrold wrote:
>
>> On Wed, 23 Apr 2014, Steve Dainard wrote:
>>
>>  I have other VM's with the same amount of snapshots without this problem.
>>> No conclusion jumping going on. More interested in what the best practice
>>> is for VM's that accumulate snapshots over time.
>>>
>>
>> For some real world context, we seem to accumulate snapshots
>> using our local approach, and are not that focused on, or
>> attentive about removing them.  The 'highwater mark' of 39, on
>> a machine that has been around since it was provisioned:
>> 2010-01-05
>>
>> [root@xxx backups]# ./count-snapshots.sh | sort -n | tail -3
>> 38 vm_64099
>> 38 vm_98036
>> 39 vm_06359
>>
>> Accumulating large numbers of snapshots seems more the
>> function of pets, than ephemeral 'cattle'
>>
>> I wrote the first paragraph without looking up the 'owners' of
>> the images. As I dereference the VM id's, all of the top ten
>> in that list turn out to be mailservers, radius servers, name
>> servers, and such, where the business unit owners chose not
>> (or neglect) to 'winnow' their herd.  There are no ephemeral
>> use units in the top ten
>>
>> -- Russ herrold
>> ___
>> Users mailing list
>> Users@ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/users
>>
>>
> please note there is a recommended limit of having no more than 500
> snapshots per block storage domain due to some LVM performance issues with
> high number of LVs. each disk/snapshot is an LV.
> NFS doesn't have this limitation.
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Ovirt snapshot failing on one VM

2014-04-23 Thread Itamar Heim

On 04/23/2014 09:57 PM, R P Herrold wrote:

On Wed, 23 Apr 2014, Steve Dainard wrote:


I have other VM's with the same amount of snapshots without this problem.
No conclusion jumping going on. More interested in what the best practice
is for VM's that accumulate snapshots over time.


For some real world context, we seem to accumulate snapshots
using our local approach, and are not that focused on, or
attentive about removing them.  The 'highwater mark' of 39, on
a machine that has been around since it was provisioned:
2010-01-05

[root@xxx backups]# ./count-snapshots.sh | sort -n | tail -3
38 vm_64099
38 vm_98036
39 vm_06359

Accumulating large numbers of snapshots seems more the
function of pets, than ephemeral 'cattle'

I wrote the first paragraph without looking up the 'owners' of
the images. As I dereference the VM id's, all of the top ten
in that list turn out to be mailservers, radius servers, name
servers, and such, where the business unit owners chose not
(or neglect) to 'winnow' their herd.  There are no ephemeral
use units in the top ten

-- Russ herrold
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users



please note there is a recommended limit of having no more than 500 
snapshots per block storage domain due to some LVM performance issues 
with high number of LVs. each disk/snapshot is an LV.

NFS doesn't have this limitation.
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] Ovirt snapshot failing on one VM

2014-04-23 Thread R P Herrold
On Wed, 23 Apr 2014, Steve Dainard wrote:

> I have other VM's with the same amount of snapshots without this problem.
> No conclusion jumping going on. More interested in what the best practice
> is for VM's that accumulate snapshots over time.

For some real world context, we seem to accumulate snapshots 
using our local approach, and are not that focused on, or 
attentive about removing them.  The 'highwater mark' of 39, on 
a machine that has been around since it was provisioned: 
2010-01-05

[root@xxx backups]# ./count-snapshots.sh | sort -n | tail -3
38 vm_64099
38 vm_98036
39 vm_06359

Accumulating large numbers of snapshots seems more the 
function of pets, than ephemeral 'cattle'

I wrote the first paragraph without looking up the 'owners' of 
the images. As I dereference the VM id's, all of the top ten 
in that list turn out to be mailservers, radius servers, name 
servers, and such, where the business unit owners chose not 
(or neglect) to 'winnow' their herd.  There are no ephemeral 
use units in the top ten

-- Russ herrold
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Ovirt snapshot failing on one VM

2014-04-23 Thread Dafna Ron
queries run on the system all the time but the only failure I 
encountered that was caused because of it was with live snapshots on a 
loaded setup in which vm running on the hsm would fail live snapshots if 
the system was loaded and the queries took a long time to come back.
However, since the error you have happens when the vm is down and only 
on that specific vm, I think it's most likely related to a failure 
somewhere in the chain.


Before committing or deleting any of the snapshots, is it possible for 
you to export the vm as is (with the snapshots) to an export domain? 
that way we know it's backed up before doing anything on the chain (and 
actually, this would be a much better way of backing up a vm rather than 
snapshots).


I don't really know what the problem in the chain is or when it 
happened, which is why I want to be cautious when continuing and delete 
of a snapshot would be better than committing it.


I can also suggest creating a new vm from any snapshot you think would 
be an important point in time for you - that case, even if there is a 
problem with the image you have a new vm with this image.


so to recap, lets try this:
1. restart vdsm and try to create a snapshot again
2. export the vm to an export domain without collapsing the snapshots
3. delete or commit the snapshots - if any fail please attach the logs. 
also, if you delete/commit, after each snapshot you can try to create a 
new snapshot to see if the issue is solved






On 04/23/2014 05:08 PM, Steve Dainard wrote:



*Steve *

On Wed, Apr 23, 2014 at 5:14 AM, Dafna Ron > wrote:


steve,
I did not say that there is a limit. there is no limit and you can
take a 1000 snapshots if you like, I simply said that I think that
it would not be would a good practice to do so.


I'm not trying to be adversarial here, but this is contradictory; if 
there's 'no limit' but 'its not good practice' and we assume that we 
want our virtual infrastructure to run smoothly, then effectively 
there is a limit we just don't know what it is.


I also did not say that this is your current problem with the vm
so you are jumping to conclusions here.


I wasn't connecting the dots between # of snapshots, and the current 
issue, I have other VM's with the same amount of snapshots without 
this problem. No conclusion jumping going on. More interested in what 
the best practice is for VM's that accumulate snapshots over time.


There is a feature slated for 3.5 
http://www.ovirt.org/Features/Live_Merge which merges snapshots on a 
running VM, so I suppose in the long run I won't have a high snapshot 
count.


i simply explained how snapshots work which is that they are
created in a chain, if there is a problem at a single point in
time it would effect the rest of the snapshots below it.


Just for clarity, such a problem would affect the snapshots 'below it' 
means after the problematic snapshot? Example: Snapshot 1,2,3,4,5. #4 
has a consistency issue, snaps 1,2,3 should be ok? I can try 
incrementally rolling back snapshots if this is the case (after vdsm 
restart suggested).


Is there any way to do a consistency check? I can imagine scheduling a 
cronjob to run through a nightly check for consistency issues, then 
roll back to an earlier snapshot to circumvent the issue.


And that we query all images under the base Image so if you have a
lot of them it would take a long time for the results to come back.


That's good to know, is this query done on new snapshot creation only? 
So over time the more snapshots I have, new snapshots will take longer 
to complete?



as for your vm, since you fail to create a snapshot on only that
vm it means that there is a problem in the current vm and it's chain.

I can see when comparing the uuid's that the pool, domain, base
image and last snapshots all exists in the rhev link.

2014-04-22 12:13:41,083 INFO
[org.ovirt.engine.core.vdsbroker.irsbroker.CreateSnapshotVDSCommand]
(pool-6-thread-49) [7ccaed5] -- createVolume parameters:
sdUUID=95b9d922-4df7-4d3b-9bca-467e2fd9d573
spUUID=9497ef2c-8368-4c92-8d61-7f318a90748f
imgGUID=466d9ae9-e46a-46f8-9f4b-964d8af0675b
size=21,474,836,480 bytes
volFormat=COW
volType=Sparse
volUUID=0b2d15e5-bf4f-4eaf-90e2-f1bd51a3a936
descr=
srcImgGUID=466d9ae9-e46a-46f8-9f4b-964d8af0675b
srcVolUUID=1a67de4b-aa1c-4436-baca-ca55726d54d7



lets see if it's possibly a cache issue - can you please restart
vdsm on the hosts?


I'll update when I have a chance to restart the services.

Thanks








On 04/22/2014 08:22 PM, Steve Dainard wrote:

All snapshots are from before failure.

That's a bit scary that there may be a 'too many snapshots'
issue. I take snapsh

Re: [ovirt-users] Ovirt snapshot failing on one VM

2014-04-23 Thread Dafna Ron

steve,
I did not say that there is a limit. there is no limit and you can take 
a 1000 snapshots if you like, I simply said that I think that it would 
not be would a good practice to do so.
I also did not say that this is your current problem with the vm so you 
are jumping to conclusions here.
i simply explained how snapshots work which is that they are created in 
a chain, if there is a problem at a single point in time it would effect 
the rest of the snapshots below it.
And that we query all images under the base Image so if you have a lot 
of them it would take a long time for the results to come back.


as for your vm, since you fail to create a snapshot on only that vm it 
means that there is a problem in the current vm and it's chain.


I can see when comparing the uuid's that the pool, domain, base image 
and last snapshots all exists in the rhev link.


2014-04-22 12:13:41,083 INFO 
[org.ovirt.engine.core.vdsbroker.irsbroker.CreateSnapshotVDSCommand] 
(pool-6-thread-49) [7ccaed5] -- createVolume parameters:

sdUUID=95b9d922-4df7-4d3b-9bca-467e2fd9d573
spUUID=9497ef2c-8368-4c92-8d61-7f318a90748f
imgGUID=466d9ae9-e46a-46f8-9f4b-964d8af0675b
size=21,474,836,480 bytes
volFormat=COW
volType=Sparse
volUUID=0b2d15e5-bf4f-4eaf-90e2-f1bd51a3a936
descr=
srcImgGUID=466d9ae9-e46a-46f8-9f4b-964d8af0675b
srcVolUUID=1a67de4b-aa1c-4436-baca-ca55726d54d7



lets see if it's possibly a cache issue - can you please restart vdsm on 
the hosts?







On 04/22/2014 08:22 PM, Steve Dainard wrote:

All snapshots are from before failure.

That's a bit scary that there may be a 'too many snapshots' issue. I 
take snapshots for point in time consistency, and without the ability 
to collapse them while the vm is running I'm not sure what the best 
option is here. What is the recommended snapshot limit? Or maybe a 
better question; whats the intended use case for snapshots in ovirt?


Export domain is currently unavailable, and without it active I can't 
disable it properly.


# ls -tl 
/rhev/data-center/9497ef2c-8368-4c92-8d61-7f318a90748f/95b9d922-4df7-4d3b-9bca-467e2fd9d573/images/466d9ae9-e46a-46f8-9f4b-964d8af0675b

total 8517740
-rw-rw. 1 vdsm kvm97583104 Apr 22 14:03 
1a67de4b-aa1c-4436-baca-ca55726d54d7
-rw-r--r--. 1 vdsm kvm 268 Apr 22 12:13 
1a67de4b-aa1c-4436-baca-ca55726d54d7.meta
-rw-r--r--. 1 vdsm kvm 272 Apr 22 01:06 
87390b64-becd-4a6f-a4fc-d27655f59b64.meta
-rw-rw. 1 vdsm kvm 1048576 Apr 22 01:04 
1a67de4b-aa1c-4436-baca-ca55726d54d7.lease
-rw-rw. 1 vdsm kvm   107413504 Apr 20 22:00 
87390b64-becd-4a6f-a4fc-d27655f59b64
-rw-rw. 1 vdsm kvm   104267776 Apr 19 22:00 
6f9fd451-6c82-4390-802c-9e23a7d89427
-rw-rw. 1 vdsm kvm 1048576 Apr 19 22:00 
87390b64-becd-4a6f-a4fc-d27655f59b64.lease
-rw-r--r--. 1 vdsm kvm 272 Apr 19 22:00 
6f9fd451-6c82-4390-802c-9e23a7d89427.meta
-rw-rw. 1 vdsm kvm   118358016 Apr 18 22:00 
c298ce3b-ec6a-4526-9971-a769f4d3d69b
-rw-rw. 1 vdsm kvm 1048576 Apr 18 22:00 
6f9fd451-6c82-4390-802c-9e23a7d89427.lease
-rw-r--r--. 1 vdsm kvm 272 Apr 18 22:00 
c298ce3b-ec6a-4526-9971-a769f4d3d69b.meta
-rw-rw. 1 vdsm kvm   120913920 Apr 17 22:00 
0ee58208-6be8-4f81-bd51-0bd4b6d5d83a
-rw-rw. 1 vdsm kvm 1048576 Apr 17 22:00 
c298ce3b-ec6a-4526-9971-a769f4d3d69b.lease
-rw-r--r--. 1 vdsm kvm 272 Apr 17 22:00 
0ee58208-6be8-4f81-bd51-0bd4b6d5d83a.meta
-rw-rw. 1 vdsm kvm   117374976 Apr 16 22:00 
9aeb973d-9a54-441e-9ce9-f4f1a233da26
-rw-rw. 1 vdsm kvm 1048576 Apr 16 22:00 
0ee58208-6be8-4f81-bd51-0bd4b6d5d83a.lease
-rw-r--r--. 1 vdsm kvm 272 Apr 16 22:00 
9aeb973d-9a54-441e-9ce9-f4f1a233da26.meta
-rw-rw. 1 vdsm kvm   110886912 Apr 15 22:00 
0eae2185-884a-44d3-9099-e952b6b7ec37
-rw-rw. 1 vdsm kvm 1048576 Apr 15 22:00 
9aeb973d-9a54-441e-9ce9-f4f1a233da26.lease
-rw-r--r--. 1 vdsm kvm 272 Apr 15 22:00 
0eae2185-884a-44d3-9099-e952b6b7ec37.meta
-rw-rw. 1 vdsm kvm 1048576 Apr 14 22:00 
0eae2185-884a-44d3-9099-e952b6b7ec37.lease
-rw-rw. 1 vdsm kvm   164560896 Apr 14 22:00 
ceffc643-b823-44b3-961e-93f3dc971886
-rw-r--r--. 1 vdsm kvm 272 Apr 14 22:00 
ceffc643-b823-44b3-961e-93f3dc971886.meta
-rw-rw. 1 vdsm kvm 1048576 Apr 13 22:00 
ceffc643-b823-44b3-961e-93f3dc971886.lease
-rw-r--r--. 1 vdsm kvm 272 Apr 13 22:00 
878fc690-ab08-489c-955b-9159f62026b1.meta
-rw-rw. 1 vdsm kvm   109182976 Apr 13 21:59 
878fc690-ab08-489c-955b-9159f62026b1
-rw-rw. 1 vdsm kvm   110297088 Apr 12 22:00 
5210eec2-a0eb-462e-95d5-7cf27db312f5
-rw-rw. 1 vdsm kvm 1048576 Apr 12 22:00 
878fc690-ab08-489c-955b-9159f62026b1.lease
-rw-r--r--. 1 vdsm kvm 272 Apr 12 22:00 
5210eec2-a0eb-462e-95d5-7cf27db312f5.meta
-rw-rw. 1 vdsm kvm76480512 Apr 11 22:00 
dcce0903-0f24-434b-9d1c-d70e3969e5e

Re: [ovirt-users] Ovirt snapshot failing on one VM

2014-04-22 Thread Steve Dainard
All snapshots are from before failure.

That's a bit scary that there may be a 'too many snapshots' issue. I take
snapshots for point in time consistency, and without the ability to
collapse them while the vm is running I'm not sure what the best option is
here. What is the recommended snapshot limit? Or maybe a better question;
whats the intended use case for snapshots in ovirt?

Export domain is currently unavailable, and without it active I can't
disable it properly.

# ls -tl
/rhev/data-center/9497ef2c-8368-4c92-8d61-7f318a90748f/95b9d922-4df7-4d3b-9bca-467e2fd9d573/images/466d9ae9-e46a-46f8-9f4b-964d8af0675b
total 8517740
-rw-rw. 1 vdsm kvm97583104 Apr 22 14:03
1a67de4b-aa1c-4436-baca-ca55726d54d7
-rw-r--r--. 1 vdsm kvm 268 Apr 22 12:13
1a67de4b-aa1c-4436-baca-ca55726d54d7.meta
-rw-r--r--. 1 vdsm kvm 272 Apr 22 01:06
87390b64-becd-4a6f-a4fc-d27655f59b64.meta
-rw-rw. 1 vdsm kvm 1048576 Apr 22 01:04
1a67de4b-aa1c-4436-baca-ca55726d54d7.lease
-rw-rw. 1 vdsm kvm   107413504 Apr 20 22:00
87390b64-becd-4a6f-a4fc-d27655f59b64
-rw-rw. 1 vdsm kvm   104267776 Apr 19 22:00
6f9fd451-6c82-4390-802c-9e23a7d89427
-rw-rw. 1 vdsm kvm 1048576 Apr 19 22:00
87390b64-becd-4a6f-a4fc-d27655f59b64.lease
-rw-r--r--. 1 vdsm kvm 272 Apr 19 22:00
6f9fd451-6c82-4390-802c-9e23a7d89427.meta
-rw-rw. 1 vdsm kvm   118358016 Apr 18 22:00
c298ce3b-ec6a-4526-9971-a769f4d3d69b
-rw-rw. 1 vdsm kvm 1048576 Apr 18 22:00
6f9fd451-6c82-4390-802c-9e23a7d89427.lease
-rw-r--r--. 1 vdsm kvm 272 Apr 18 22:00
c298ce3b-ec6a-4526-9971-a769f4d3d69b.meta
-rw-rw. 1 vdsm kvm   120913920 Apr 17 22:00
0ee58208-6be8-4f81-bd51-0bd4b6d5d83a
-rw-rw. 1 vdsm kvm 1048576 Apr 17 22:00
c298ce3b-ec6a-4526-9971-a769f4d3d69b.lease
-rw-r--r--. 1 vdsm kvm 272 Apr 17 22:00
0ee58208-6be8-4f81-bd51-0bd4b6d5d83a.meta
-rw-rw. 1 vdsm kvm   117374976 Apr 16 22:00
9aeb973d-9a54-441e-9ce9-f4f1a233da26
-rw-rw. 1 vdsm kvm 1048576 Apr 16 22:00
0ee58208-6be8-4f81-bd51-0bd4b6d5d83a.lease
-rw-r--r--. 1 vdsm kvm 272 Apr 16 22:00
9aeb973d-9a54-441e-9ce9-f4f1a233da26.meta
-rw-rw. 1 vdsm kvm   110886912 Apr 15 22:00
0eae2185-884a-44d3-9099-e952b6b7ec37
-rw-rw. 1 vdsm kvm 1048576 Apr 15 22:00
9aeb973d-9a54-441e-9ce9-f4f1a233da26.lease
-rw-r--r--. 1 vdsm kvm 272 Apr 15 22:00
0eae2185-884a-44d3-9099-e952b6b7ec37.meta
-rw-rw. 1 vdsm kvm 1048576 Apr 14 22:00
0eae2185-884a-44d3-9099-e952b6b7ec37.lease
-rw-rw. 1 vdsm kvm   164560896 Apr 14 22:00
ceffc643-b823-44b3-961e-93f3dc971886
-rw-r--r--. 1 vdsm kvm 272 Apr 14 22:00
ceffc643-b823-44b3-961e-93f3dc971886.meta
-rw-rw. 1 vdsm kvm 1048576 Apr 13 22:00
ceffc643-b823-44b3-961e-93f3dc971886.lease
-rw-r--r--. 1 vdsm kvm 272 Apr 13 22:00
878fc690-ab08-489c-955b-9159f62026b1.meta
-rw-rw. 1 vdsm kvm   109182976 Apr 13 21:59
878fc690-ab08-489c-955b-9159f62026b1
-rw-rw. 1 vdsm kvm   110297088 Apr 12 22:00
5210eec2-a0eb-462e-95d5-7cf27db312f5
-rw-rw. 1 vdsm kvm 1048576 Apr 12 22:00
878fc690-ab08-489c-955b-9159f62026b1.lease
-rw-r--r--. 1 vdsm kvm 272 Apr 12 22:00
5210eec2-a0eb-462e-95d5-7cf27db312f5.meta
-rw-rw. 1 vdsm kvm76480512 Apr 11 22:00
dcce0903-0f24-434b-9d1c-d70e3969e5ea
-rw-rw. 1 vdsm kvm 1048576 Apr 11 22:00
5210eec2-a0eb-462e-95d5-7cf27db312f5.lease
-rw-r--r--. 1 vdsm kvm 272 Apr 11 22:00
dcce0903-0f24-434b-9d1c-d70e3969e5ea.meta
-rw-rw. 1 vdsm kvm 1048576 Apr 11 12:34
dcce0903-0f24-434b-9d1c-d70e3969e5ea.lease
-rw-r--r--. 1 vdsm kvm 272 Apr 11 12:34
d3a1c505-8f6a-4c2b-97b7-764cd5baea47.meta
-rw-rw. 1 vdsm kvm   20824 Apr 11 12:33
d3a1c505-8f6a-4c2b-97b7-764cd5baea47
-rw-rw. 1 vdsm kvm14614528 Apr 10 16:12
638c2164-2edc-4294-ac99-c51963140940
-rw-rw. 1 vdsm kvm 1048576 Apr 10 16:12
d3a1c505-8f6a-4c2b-97b7-764cd5baea47.lease
-rw-r--r--. 1 vdsm kvm 272 Apr 10 16:12
638c2164-2edc-4294-ac99-c51963140940.meta
-rw-rw. 1 vdsm kvm12779520 Apr 10 16:06
f8f1f164-c0d9-4716-9ab3-9131179a79bd
-rw-rw. 1 vdsm kvm 1048576 Apr 10 16:05
638c2164-2edc-4294-ac99-c51963140940.lease
-rw-r--r--. 1 vdsm kvm 272 Apr 10 16:05
f8f1f164-c0d9-4716-9ab3-9131179a79bd.meta
-rw-rw. 1 vdsm kvm92995584 Apr 10 16:00
f9b14795-a26c-4edb-ae34-22361531a0a1
-rw-rw. 1 vdsm kvm 1048576 Apr 10 16:00
f8f1f164-c0d9-4716-9ab3-9131179a79bd.lease
-rw-r--r--. 1 vdsm kvm 272 Apr 10 16:00
f9b14795-a26c-4edb-ae34-22361531a0a1.meta
-rw-rw. 1 vdsm kvm30015488 Apr 10 14:57
39cbf947-f084-4e75-8d6b-b3e5c32b82d6
-rw-rw. 1 vdsm kvm 1048576 Apr 10 14:57
f9b14795-a26c-4edb-ae34-22361531a0a1.lease
-rw-r--r--. 1 vdsm kvm 272 Apr 10 14:57
39cbf947-f084-4e75-8d6b-b3e5c32b82d6.meta
-rw-rw. 1 vdsm kvm19267584 Apr 10 14:34
3ece1489-9bff-4223-ab97-e45135106222
-rw-rw. 1 vdsm kvm 1048576 Apr 10 14:34
39cbf947-f084-4e75-8d6b-b3e5c32b82d6.lease
-rw-r--r--. 1 vdsm kvm 

Re: [ovirt-users] Ovirt snapshot failing on one VM

2014-04-22 Thread Dafna Ron

it's the same error:

c1d7c4e-392b-4a62-9836-3add1360a46d::DEBUG::2014-04-22 
12:13:44,340::volume::1058::Storage.Misc.excCmd::(createVolume) FAILED: 
 = 
'/rhev/data-center/9497ef2c-8368-4c92-8d61-7f318a90748f/95b9d922-4df7-4d3b-9bca-467e2fd9d573/images/4
66d9ae9-e46a-46f8-9f4b-964d8af0675b/0b2d15e5-bf4f-4eaf-90e2-f1bd51a3a936: error 
while creating qcow2: No such file or directory\n';  = 1



were these 23 snapshots created any way each time we fail to create the 
snapshot or are these older snapshots which you actually created before 
the failure?


at this point my main theory is that somewhere along the line you had 
some sort of failure in your storage and from that time each snapshot 
you create will fail.
if the snapshots are created during the failure can you please delete 
the snapshots you do not need and try again?


There should not be a limit on how many snapshots you can have since 
it's only a link changing the image the vm should boot from.
Having said that, it's not ideal to have that many snapshots and can 
probably lead to unexpected results so I would not recommend having that 
many snapshots on a single vm :)


for example, my second theory would be that because we have so many 
snapshots we have some sort of race where part of the createVolume 
command expects some result from a query run before the create itself 
and because there are so many snapshots there is "no such file" on the 
volume because it's too far up the list.


can you also run: ls -l 
/rhev/data-center/9497ef2c-8368-4c92-8d61-7f318a90748f/95b9d922-4df7-4d3b-9bca-467e2fd9d573/images/466d9ae9-e46a-46f8-9f4b-964d8af0675b


lets see what images are listed under that vm.

btw, you know that your export domain is getting 
StorageDomainDoesNotExist in the vdsm log? is that domain in up state? 
can you try to deactivate the export domain?


Thanks,

Dafna




On 04/22/2014 05:20 PM, Steve Dainard wrote:

Ominous..

23 snapshots. Is there an upper limit?

Offline snapshot fails as well. Both logs attached again (snapshot 
attempted at 12:13 EST).


*Steve *

On Tue, Apr 22, 2014 at 11:20 AM, Dafna Ron > wrote:


are you able to take an offline snapshot? (while the vm is down)
how many snapshots do you have on this vm?


On 04/22/2014 04:19 PM, Steve Dainard wrote:

No alert in web ui, I restarted the VM yesterday just in case,
no change. I also restored an earlier snapshot and tried to
re-snapshot, same result.

*Steve
*



On Tue, Apr 22, 2014 at 10:57 AM, Dafna Ron mailto:d...@redhat.com> >> wrote:

This is the actual problem:

bf025a73-eeeb-4ac5-b8a9-32afa4ae482e::DEBUG::2014-04-22
   
10:21:49,374::volume::1058::Storage.Misc.excCmd::(createVolume)

FAILED:  =
   
'/rhev/data-center/9497ef2c-8368-4c92-8d61-7f318a90748f/95b9d922-4df7-4d3b-9bca-467e2fd9d573/images/4
   
66d9ae9-e46a-46f8-9f4b-964d8af0675b/87efa937-b31f-4bb1-aee1-0ee14a0dc6fb:

error while creating qcow2: No such file or directory\n';
 = 1

from that you see the actual failure:

bf025a73-eeeb-4ac5-b8a9-32afa4ae482e::ERROR::2014-04-22
10:21:49,392::volume::286::Storage.Volume::(clone)
Volume.clone:
can't clone:
   
/rhev/data-center/9497ef2c-8368-4c92-8d61-7f318a90748f/95b9d922-4df7-4d3b-9bca-467e2fd9d573/images/466d
   
9ae9-e46a-46f8-9f4b-964d8af0675b/1a67de4b-aa1c-4436-baca-ca55726d54d7

to
   
/rhev/data-center/9497ef2c-8368-4c92-8d61-7f318a90748f/95b9d922-4df7-4d3b-9bca-467e2fd9d573/images/466d9ae9-e46a-46f8-9f4b-964d8af0675b/87efa937-b31f-4bb1-aee1-0ee1

4a0dc6fb
bf025a73-eeeb-4ac5-b8a9-32afa4ae482e::ERROR::2014-04-22
10:21:49,392::volume::508::Storage.Volume::(create)
Unexpected error
Traceback (most recent call last):
  File "/usr/share/vdsm/storage/volume.py", line 466, in
create
srcVolUUID, imgPath, volPath)
  File "/usr/share/vdsm/storage/fileVolume.py", line 160,
in _create
volParent.clone(imgPath, volUUID, volFormat, preallocate)
  File "/usr/share/vdsm/storage/volume.py", line 287, in clone
raise se.CannotCloneVolume(self.volumePath, dst_path,
str(e))
CannotCloneVolume: Cannot clone volume:
   
'src=/rhev/data-center/9497ef2c-8368-4c92-8d61-7f318a90748f/95b9d922-4df7-4d3b-9bca-467e2fd9d573/images/466d9ae9-e46a-46f8-9f4b-964d8af0675b/1a67de4b-aa1c-4436-baca-ca55726d54d7,

dst=/rhev/data-cen
   
ter/9497ef2c-8368-4c92-8d61-7f318a90748f/95b9d922-4df7-4d3b-9bca-467e2fd9d573/images/466d9ae9-e46a-46f8-9f4b-964d8af0675b/87efa937-b31f-4bb1-aee1-0ee14a0dc6fb:

Error creatin

Re: [ovirt-users] Ovirt snapshot failing on one VM

2014-04-22 Thread Dafna Ron

are you able to take an offline snapshot? (while the vm is down)
how many snapshots do you have on this vm?

On 04/22/2014 04:19 PM, Steve Dainard wrote:
No alert in web ui, I restarted the VM yesterday just in case, no 
change. I also restored an earlier snapshot and tried to re-snapshot, 
same result.


*Steve
*


On Tue, Apr 22, 2014 at 10:57 AM, Dafna Ron > wrote:


This is the actual problem:

bf025a73-eeeb-4ac5-b8a9-32afa4ae482e::DEBUG::2014-04-22
10:21:49,374::volume::1058::Storage.Misc.excCmd::(createVolume)
FAILED:  =

'/rhev/data-center/9497ef2c-8368-4c92-8d61-7f318a90748f/95b9d922-4df7-4d3b-9bca-467e2fd9d573/images/4
66d9ae9-e46a-46f8-9f4b-964d8af0675b/87efa937-b31f-4bb1-aee1-0ee14a0dc6fb:
error while creating qcow2: No such file or directory\n';  = 1

from that you see the actual failure:

bf025a73-eeeb-4ac5-b8a9-32afa4ae482e::ERROR::2014-04-22
10:21:49,392::volume::286::Storage.Volume::(clone) Volume.clone:
can't clone:

/rhev/data-center/9497ef2c-8368-4c92-8d61-7f318a90748f/95b9d922-4df7-4d3b-9bca-467e2fd9d573/images/466d
9ae9-e46a-46f8-9f4b-964d8af0675b/1a67de4b-aa1c-4436-baca-ca55726d54d7
to

/rhev/data-center/9497ef2c-8368-4c92-8d61-7f318a90748f/95b9d922-4df7-4d3b-9bca-467e2fd9d573/images/466d9ae9-e46a-46f8-9f4b-964d8af0675b/87efa937-b31f-4bb1-aee1-0ee1
4a0dc6fb
bf025a73-eeeb-4ac5-b8a9-32afa4ae482e::ERROR::2014-04-22
10:21:49,392::volume::508::Storage.Volume::(create) Unexpected error
Traceback (most recent call last):
  File "/usr/share/vdsm/storage/volume.py", line 466, in create
srcVolUUID, imgPath, volPath)
  File "/usr/share/vdsm/storage/fileVolume.py", line 160, in _create
volParent.clone(imgPath, volUUID, volFormat, preallocate)
  File "/usr/share/vdsm/storage/volume.py", line 287, in clone
raise se.CannotCloneVolume(self.volumePath, dst_path, str(e))
CannotCloneVolume: Cannot clone volume:

'src=/rhev/data-center/9497ef2c-8368-4c92-8d61-7f318a90748f/95b9d922-4df7-4d3b-9bca-467e2fd9d573/images/466d9ae9-e46a-46f8-9f4b-964d8af0675b/1a67de4b-aa1c-4436-baca-ca55726d54d7,
dst=/rhev/data-cen

ter/9497ef2c-8368-4c92-8d61-7f318a90748f/95b9d922-4df7-4d3b-9bca-467e2fd9d573/images/466d9ae9-e46a-46f8-9f4b-964d8af0675b/87efa937-b31f-4bb1-aee1-0ee14a0dc6fb:
Error creating a new volume: (["Formatting
\'/rhev/data-center/9497ef2c-8368-

4c92-8d61-7f318a90748f/95b9d922-4df7-4d3b-9bca-467e2fd9d573/images/466d9ae9-e46a-46f8-9f4b-964d8af0675b/87efa937-b31f-4bb1-aee1-0ee14a0dc6fb\',
fmt=qcow2 size=21474836480
backing_file=\'../466d9ae9-e46a-46f8-9f4b-964d8af0675b/1a67de4b-aa
1c-4436-baca-ca55726d54d7\' backing_fmt=\'qcow2\' encryption=off
cluster_size=65536 "],)'


do you have any alert in the webadmin to restart the vm?

Dafna


On 04/22/2014 03:31 PM, Steve Dainard wrote:

Sorry for the confusion.

I attempted to take a live snapshot of a running VM. After
that failed, I migrated the VM to another host, and attempted
the live snapshot again without success, eliminating a single
host as the cause of failure.

Ovirt is 3.3.4, storage domain is gluster 3.4.2.1, OS is
CentOS 6.5.

Package versions:
libvirt-0.10.2-29.el6_5.5.x86_64
libvirt-lock-sanlock-0.10.2-29.el6_5.5.x86_64
qemu-img-rhev-0.12.1.2-2.415.el6.nux.3.x86_64
qemu-kvm-rhev-0.12.1.2-2.415.el6.nux.3.x86_64
qemu-kvm-rhev-tools-0.12.1.2-2.415.el6.nux.3.x86_64
vdsm-4.13.3-4.el6.x86_64
vdsm-gluster-4.13.3-4.el6.noarch


I made another live snapshot attempt at 10:21 EST today, full
vdsm.log attached, and a truncated engine.log.

Thanks,

*Steve
*



On Tue, Apr 22, 2014 at 9:48 AM, Dafna Ron mailto:d...@redhat.com> >> wrote:

please explain the flow of what you are trying to do, are you
trying to live migrate the disk (from one storage to
another), are
you trying to migrate the vm and after vm migration is
finished
you try to take a live snapshot of the vm? or are you
trying to
take a live snapshot of the vm during a vm migration from
host1 to
host2?

Please attach full vdsm logs from any host you are using
(if you
are trying to migrate the vm from host1 to host2) + please
attach
engine log.

Also, what is the vdsm, libvirt and qemu versions, what ovirt
version are you using and what is the storage you are using?

Thanks,

Dafna




On 04/22/2014 02:12 PM, Steve Dainard wrote:

I've attempted migrating the vm to another host and
taking a
snapshot, but I get this error:

6efd

Re: [ovirt-users] Ovirt snapshot failing on one VM

2014-04-22 Thread Steve Dainard
No alert in web ui, I restarted the VM yesterday just in case, no change. I
also restored an earlier snapshot and tried to re-snapshot, same result.


*Steve *


On Tue, Apr 22, 2014 at 10:57 AM, Dafna Ron  wrote:

> This is the actual problem:
>
> bf025a73-eeeb-4ac5-b8a9-32afa4ae482e::DEBUG::2014-04-22
> 10:21:49,374::volume::1058::Storage.Misc.excCmd::(createVolume) FAILED:
>  = '/rhev/data-center/9497ef2c-8368-4c92-8d61-7f318a90748f/
> 95b9d922-4df7-4d3b-9bca-467e2fd9d573/images/4
> 66d9ae9-e46a-46f8-9f4b-964d8af0675b/87efa937-b31f-4bb1-aee1-0ee14a0dc6fb:
> error while creating qcow2: No such file or directory\n';  = 1
>
> from that you see the actual failure:
>
> bf025a73-eeeb-4ac5-b8a9-32afa4ae482e::ERROR::2014-04-22
> 10:21:49,392::volume::286::Storage.Volume::(clone) Volume.clone: can't
> clone: /rhev/data-center/9497ef2c-8368-4c92-8d61-7f318a90748f/
> 95b9d922-4df7-4d3b-9bca-467e2fd9d573/images/466d
> 9ae9-e46a-46f8-9f4b-964d8af0675b/1a67de4b-aa1c-4436-baca-ca55726d54d7 to
> /rhev/data-center/9497ef2c-8368-4c92-8d61-7f318a90748f/
> 95b9d922-4df7-4d3b-9bca-467e2fd9d573/images/466d9ae9-
> e46a-46f8-9f4b-964d8af0675b/87efa937-b31f-4bb1-aee1-0ee1
> 4a0dc6fb
> bf025a73-eeeb-4ac5-b8a9-32afa4ae482e::ERROR::2014-04-22
> 10:21:49,392::volume::508::Storage.Volume::(create) Unexpected error
> Traceback (most recent call last):
>   File "/usr/share/vdsm/storage/volume.py", line 466, in create
> srcVolUUID, imgPath, volPath)
>   File "/usr/share/vdsm/storage/fileVolume.py", line 160, in _create
> volParent.clone(imgPath, volUUID, volFormat, preallocate)
>   File "/usr/share/vdsm/storage/volume.py", line 287, in clone
> raise se.CannotCloneVolume(self.volumePath, dst_path, str(e))
> CannotCloneVolume: Cannot clone volume: 'src=/rhev/data-center/
> 9497ef2c-8368-4c92-8d61-7f318a90748f/95b9d922-4df7-
> 4d3b-9bca-467e2fd9d573/images/466d9ae9-e46a-46f8-9f4b-
> 964d8af0675b/1a67de4b-aa1c-4436-baca-ca55726d54d7, dst=/rhev/data-cen
> ter/9497ef2c-8368-4c92-8d61-7f318a90748f/95b9d922-4df7-
> 4d3b-9bca-467e2fd9d573/images/466d9ae9-e46a-46f8-9f4b-
> 964d8af0675b/87efa937-b31f-4bb1-aee1-0ee14a0dc6fb: Error creating a new
> volume: (["Formatting \'/rhev/data-center/9497ef2c-8368-
> 4c92-8d61-7f318a90748f/95b9d922-4df7-4d3b-9bca-
> 467e2fd9d573/images/466d9ae9-e46a-46f8-9f4b-964d8af0675b/
> 87efa937-b31f-4bb1-aee1-0ee14a0dc6fb\', fmt=qcow2 size=21474836480
> backing_file=\'../466d9ae9-e46a-46f8-9f4b-964d8af0675b/1a67de4b-aa
> 1c-4436-baca-ca55726d54d7\' backing_fmt=\'qcow2\' encryption=off
> cluster_size=65536 "],)'
>
>
> do you have any alert in the webadmin to restart the vm?
>
> Dafna
>
>
> On 04/22/2014 03:31 PM, Steve Dainard wrote:
>
>> Sorry for the confusion.
>>
>> I attempted to take a live snapshot of a running VM. After that failed, I
>> migrated the VM to another host, and attempted the live snapshot again
>> without success, eliminating a single host as the cause of failure.
>>
>> Ovirt is 3.3.4, storage domain is gluster 3.4.2.1, OS is CentOS 6.5.
>>
>> Package versions:
>> libvirt-0.10.2-29.el6_5.5.x86_64
>> libvirt-lock-sanlock-0.10.2-29.el6_5.5.x86_64
>> qemu-img-rhev-0.12.1.2-2.415.el6.nux.3.x86_64
>> qemu-kvm-rhev-0.12.1.2-2.415.el6.nux.3.x86_64
>> qemu-kvm-rhev-tools-0.12.1.2-2.415.el6.nux.3.x86_64
>> vdsm-4.13.3-4.el6.x86_64
>> vdsm-gluster-4.13.3-4.el6.noarch
>>
>>
>> I made another live snapshot attempt at 10:21 EST today, full vdsm.log
>> attached, and a truncated engine.log.
>>
>> Thanks,
>>
>> *Steve
>> *
>>
>>
>>
>> On Tue, Apr 22, 2014 at 9:48 AM, Dafna Ron > d...@redhat.com>> wrote:
>>
>> please explain the flow of what you are trying to do, are you
>> trying to live migrate the disk (from one storage to another), are
>> you trying to migrate the vm and after vm migration is finished
>> you try to take a live snapshot of the vm? or are you trying to
>> take a live snapshot of the vm during a vm migration from host1 to
>> host2?
>>
>> Please attach full vdsm logs from any host you are using (if you
>> are trying to migrate the vm from host1 to host2) + please attach
>> engine log.
>>
>> Also, what is the vdsm, libvirt and qemu versions, what ovirt
>> version are you using and what is the storage you are using?
>>
>> Thanks,
>>
>> Dafna
>>
>>
>>
>>
>> On 04/22/2014 02:12 PM, Steve Dainard wrote:
>>
>> I've attempted migrating the vm to another host and taking a
>> snapshot, but I get this error:
>>
>> 6efd33f4-984c-4513-b5e6-fffdca2e983b::ERROR::2014-04-22
>> 01:09:37,296::volume::286::Storage.Volume::(clone)
>> Volume.clone: can't clone:
>> /rhev/data-center/9497ef2c-8368-4c92-8d61-7f318a90748f/
>> 95b9d922-4df7-4d3b-9bca-467e2fd9d573/images/466d9ae9-
>> e46a-46f8-9f4b-964d8af0675b/1a67de4b-aa1c-4436-baca-ca55726d54d7
>> to
>> /rhev/data-center/9497ef2c-8368-4c92-8d61-7f318a90748f/
>> 95b9d922-4df7-4d3b-9bca-467e2fd9d573/images/466d9ae9-
>> e46a-46f8-9f4b-964d8af0675b

Re: [ovirt-users] Ovirt snapshot failing on one VM

2014-04-22 Thread Dafna Ron

This is the actual problem:

bf025a73-eeeb-4ac5-b8a9-32afa4ae482e::DEBUG::2014-04-22 
10:21:49,374::volume::1058::Storage.Misc.excCmd::(createVolume) FAILED: 
 = 
'/rhev/data-center/9497ef2c-8368-4c92-8d61-7f318a90748f/95b9d922-4df7-4d3b-9bca-467e2fd9d573/images/4
66d9ae9-e46a-46f8-9f4b-964d8af0675b/87efa937-b31f-4bb1-aee1-0ee14a0dc6fb: error 
while creating qcow2: No such file or directory\n';  = 1


from that you see the actual failure:

bf025a73-eeeb-4ac5-b8a9-32afa4ae482e::ERROR::2014-04-22 
10:21:49,392::volume::286::Storage.Volume::(clone) Volume.clone: can't 
clone: 
/rhev/data-center/9497ef2c-8368-4c92-8d61-7f318a90748f/95b9d922-4df7-4d3b-9bca-467e2fd9d573/images/466d
9ae9-e46a-46f8-9f4b-964d8af0675b/1a67de4b-aa1c-4436-baca-ca55726d54d7 to 
/rhev/data-center/9497ef2c-8368-4c92-8d61-7f318a90748f/95b9d922-4df7-4d3b-9bca-467e2fd9d573/images/466d9ae9-e46a-46f8-9f4b-964d8af0675b/87efa937-b31f-4bb1-aee1-0ee1

4a0dc6fb
bf025a73-eeeb-4ac5-b8a9-32afa4ae482e::ERROR::2014-04-22 
10:21:49,392::volume::508::Storage.Volume::(create) Unexpected error

Traceback (most recent call last):
  File "/usr/share/vdsm/storage/volume.py", line 466, in create
srcVolUUID, imgPath, volPath)
  File "/usr/share/vdsm/storage/fileVolume.py", line 160, in _create
volParent.clone(imgPath, volUUID, volFormat, preallocate)
  File "/usr/share/vdsm/storage/volume.py", line 287, in clone
raise se.CannotCloneVolume(self.volumePath, dst_path, str(e))
CannotCloneVolume: Cannot clone volume: 
'src=/rhev/data-center/9497ef2c-8368-4c92-8d61-7f318a90748f/95b9d922-4df7-4d3b-9bca-467e2fd9d573/images/466d9ae9-e46a-46f8-9f4b-964d8af0675b/1a67de4b-aa1c-4436-baca-ca55726d54d7, 
dst=/rhev/data-cen
ter/9497ef2c-8368-4c92-8d61-7f318a90748f/95b9d922-4df7-4d3b-9bca-467e2fd9d573/images/466d9ae9-e46a-46f8-9f4b-964d8af0675b/87efa937-b31f-4bb1-aee1-0ee14a0dc6fb: 
Error creating a new volume: (["Formatting 
\'/rhev/data-center/9497ef2c-8368-
4c92-8d61-7f318a90748f/95b9d922-4df7-4d3b-9bca-467e2fd9d573/images/466d9ae9-e46a-46f8-9f4b-964d8af0675b/87efa937-b31f-4bb1-aee1-0ee14a0dc6fb\', 
fmt=qcow2 size=21474836480 
backing_file=\'../466d9ae9-e46a-46f8-9f4b-964d8af0675b/1a67de4b-aa
1c-4436-baca-ca55726d54d7\' backing_fmt=\'qcow2\' encryption=off 
cluster_size=65536 "],)'



do you have any alert in the webadmin to restart the vm?

Dafna

On 04/22/2014 03:31 PM, Steve Dainard wrote:

Sorry for the confusion.

I attempted to take a live snapshot of a running VM. After that 
failed, I migrated the VM to another host, and attempted the live 
snapshot again without success, eliminating a single host as the cause 
of failure.


Ovirt is 3.3.4, storage domain is gluster 3.4.2.1, OS is CentOS 6.5.

Package versions:
libvirt-0.10.2-29.el6_5.5.x86_64
libvirt-lock-sanlock-0.10.2-29.el6_5.5.x86_64
qemu-img-rhev-0.12.1.2-2.415.el6.nux.3.x86_64
qemu-kvm-rhev-0.12.1.2-2.415.el6.nux.3.x86_64
qemu-kvm-rhev-tools-0.12.1.2-2.415.el6.nux.3.x86_64
vdsm-4.13.3-4.el6.x86_64
vdsm-gluster-4.13.3-4.el6.noarch


I made another live snapshot attempt at 10:21 EST today, full vdsm.log 
attached, and a truncated engine.log.


Thanks,

*Steve
*


On Tue, Apr 22, 2014 at 9:48 AM, Dafna Ron > wrote:


please explain the flow of what you are trying to do, are you
trying to live migrate the disk (from one storage to another), are
you trying to migrate the vm and after vm migration is finished
you try to take a live snapshot of the vm? or are you trying to
take a live snapshot of the vm during a vm migration from host1 to
host2?

Please attach full vdsm logs from any host you are using (if you
are trying to migrate the vm from host1 to host2) + please attach
engine log.

Also, what is the vdsm, libvirt and qemu versions, what ovirt
version are you using and what is the storage you are using?

Thanks,

Dafna




On 04/22/2014 02:12 PM, Steve Dainard wrote:

I've attempted migrating the vm to another host and taking a
snapshot, but I get this error:

6efd33f4-984c-4513-b5e6-fffdca2e983b::ERROR::2014-04-22
01:09:37,296::volume::286::Storage.Volume::(clone)
Volume.clone: can't clone:

/rhev/data-center/9497ef2c-8368-4c92-8d61-7f318a90748f/95b9d922-4df7-4d3b-9bca-467e2fd9d573/images/466d9ae9-e46a-46f8-9f4b-964d8af0675b/1a67de4b-aa1c-4436-baca-ca55726d54d7
to

/rhev/data-center/9497ef2c-8368-4c92-8d61-7f318a90748f/95b9d922-4df7-4d3b-9bca-467e2fd9d573/images/466d9ae9-e46a-46f8-9f4b-964d8af0675b/b230596f-97bc-4532-ba57-5654fa9c6c51

A bit more of the vdsm log is attached.

Other vm's are snapshotting without issue.



Any help appreciated,

*Steve
*


___
Users mailing list
Users@ovirt.org 
http://lists.ovirt.org/mailman/listinfo/users



-- 
Dafna Ron






--
Dafna Ron
__

Re: [ovirt-users] Ovirt snapshot failing on one VM

2014-04-22 Thread Aharon Canan
hi 

please check https://bugzilla.redhat.com/show_bug.cgi?id=1017289 
I know that snapshot on gluster domain has few issues... 



Regards, 
__ 
Aharon Canan 


- Original Message -

From: "Steve Dainard"  
To: d...@redhat.com 
Cc: "users"  
Sent: Tuesday, April 22, 2014 5:31:18 PM 
Subject: Re: [ovirt-users] Ovirt snapshot failing on one VM 

Sorry for the confusion. 

I attempted to take a live snapshot of a running VM. After that failed, I 
migrated the VM to another host, and attempted the live snapshot again without 
success, eliminating a single host as the cause of failure. 

Ovirt is 3.3.4, storage domain is gluster 3.4.2.1, OS is CentOS 6.5. 

Package versions: 
libvirt-0.10.2-29.el6_5.5.x86_64 
libvirt-lock-sanlock-0.10.2-29.el6_5.5.x86_64 
qemu-img-rhev-0.12.1.2-2.415.el6.nux.3.x86_64 
qemu-kvm-rhev-0.12.1.2-2.415.el6.nux.3.x86_64 
qemu-kvm-rhev-tools-0.12.1.2-2.415.el6.nux.3.x86_64 
vdsm-4.13.3-4.el6.x86_64 
vdsm-gluster-4.13.3-4.el6.noarch 


I made another live snapshot attempt at 10:21 EST today, full vdsm.log 
attached, and a truncated engine.log. 

Thanks, 

Steve 


On Tue, Apr 22, 2014 at 9:48 AM, Dafna Ron < d...@redhat.com > wrote: 


please explain the flow of what you are trying to do, are you trying to live 
migrate the disk (from one storage to another), are you trying to migrate the 
vm and after vm migration is finished you try to take a live snapshot of the 
vm? or are you trying to take a live snapshot of the vm during a vm migration 
from host1 to host2? 

Please attach full vdsm logs from any host you are using (if you are trying to 
migrate the vm from host1 to host2) + please attach engine log. 

Also, what is the vdsm, libvirt and qemu versions, what ovirt version are you 
using and what is the storage you are using? 

Thanks, 

Dafna 




On 04/22/2014 02:12 PM, Steve Dainard wrote: 



I've attempted migrating the vm to another host and taking a snapshot, but I 
get this error: 

6efd33f4-984c-4513-b5e6- fffdca2e983b::ERROR::2014-04- 22 
01:09:37,296::volume::286:: Storage.Volume::(clone) Volume.clone: can't clone: 
/rhev/data-center/9497ef2c- 8368-4c92-8d61-7f318a90748f/ 
95b9d922-4df7-4d3b-9bca- 467e2fd9d573/images/466d9ae9- 
e46a-46f8-9f4b-964d8af0675b/ 1a67de4b-aa1c-4436-baca- ca55726d54d7 to 
/rhev/data-center/9497ef2c- 8368-4c92-8d61-7f318a90748f/ 
95b9d922-4df7-4d3b-9bca- 467e2fd9d573/images/466d9ae9- 
e46a-46f8-9f4b-964d8af0675b/ b230596f-97bc-4532-ba57- 5654fa9c6c51 

A bit more of the vdsm log is attached. 

Other vm's are snapshotting without issue. 



Any help appreciated, 

*Steve 
* 


__ _ 
Users mailing list 
Users@ovirt.org 
http://lists.ovirt.org/ mailman/listinfo/users 




-- 
Dafna Ron 





___ 
Users mailing list 
Users@ovirt.org 
http://lists.ovirt.org/mailman/listinfo/users 

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Ovirt snapshot failing on one VM

2014-04-22 Thread Dafna Ron
please explain the flow of what you are trying to do, are you trying to 
live migrate the disk (from one storage to another), are you trying to 
migrate the vm and after vm migration is finished you try to take a live 
snapshot of the vm? or are you trying to take a live snapshot of the vm 
during a vm migration from host1 to host2?


Please attach full vdsm logs from any host you are using (if you are 
trying to migrate the vm from host1 to host2) + please attach engine log.


Also, what is the vdsm, libvirt and qemu versions, what ovirt version 
are you using and what is the storage you are using?


Thanks,

Dafna



On 04/22/2014 02:12 PM, Steve Dainard wrote:
I've attempted migrating the vm to another host and taking a snapshot, 
but I get this error:


6efd33f4-984c-4513-b5e6-fffdca2e983b::ERROR::2014-04-22 
01:09:37,296::volume::286::Storage.Volume::(clone) Volume.clone: can't 
clone: 
/rhev/data-center/9497ef2c-8368-4c92-8d61-7f318a90748f/95b9d922-4df7-4d3b-9bca-467e2fd9d573/images/466d9ae9-e46a-46f8-9f4b-964d8af0675b/1a67de4b-aa1c-4436-baca-ca55726d54d7 
to 
/rhev/data-center/9497ef2c-8368-4c92-8d61-7f318a90748f/95b9d922-4df7-4d3b-9bca-467e2fd9d573/images/466d9ae9-e46a-46f8-9f4b-964d8af0675b/b230596f-97bc-4532-ba57-5654fa9c6c51


A bit more of the vdsm log is attached.

Other vm's are snapshotting without issue.



Any help appreciated,

*Steve
*


___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users



--
Dafna Ron
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users