On Wed, Feb 22, 2017 at 10:32 AM, Nir Soffer <nsof...@redhat.com> wrote:
> On Wed, Feb 22, 2017 at 10:31 AM, Nir Soffer <nsof...@redhat.com> wrote:
>> On Mon, Feb 20, 2017 at 4:49 PM, Gianluca Cecchi
>> <gianluca.cec...@gmail.com> wrote:
>>> Hello,
>>> I'm trying to move a disk from one storage domain A to another B in oVirt
>>> 4.1
>>> The corresponding VM is powered on in the mean time
>>>
>>> When executing the action, there was already in place a disk move from
>>> storage domain C to A (this move was for a disk of a powered off VM and then
>>> completed ok)
>>> I got this in events of webadmin gui for the failed move A -> B:
>>>
>>> Feb 20, 2017 2:42:00 PM Failed to complete snapshot 'Auto-generated for Live
>>> Storage Migration' creation for VM 'dbatest6'.
>>> Feb 20, 2017 2:40:51 PM VDSM ovmsrv06 command HSMGetAllTasksStatusesVDS
>>> failed: Error creating a new volume
>>> Feb 20, 2017 2:40:51 PM Snapshot 'Auto-generated for Live Storage Migration'
>>> creation for VM 'dbatest6' was initiated by admin@internal-authz.
>>>
>>>
>>> And in relevant vdsm.log of referred host ovmsrv06
>>>
>>> 2017-02-20 14:41:44,899 ERROR (tasks/8) [storage.Volume] Unexpected error
>>> (volume:1087)
>>> Traceback (most recent call last):
>>>   File "/usr/share/vdsm/storage/volume.py", line 1081, in create
>>>     cls.newVolumeLease(metaId, sdUUID, volUUID)
>>>   File "/usr/share/vdsm/storage/volume.py", line 1361, in newVolumeLease
>>>     return cls.manifestClass.newVolumeLease(metaId, sdUUID, volUUID)
>>>   File "/usr/share/vdsm/storage/blockVolume.py", line 310, in newVolumeLease
>>>     sanlock.init_resource(sdUUID, volUUID, [(leasePath, leaseOffset)])
>>> SanlockException: (-202, 'Sanlock resource init failure', 'Sanlock
>>> exception')
>>
>> This means that sanlock could not initialize a lease in the new volume 
>> created
>> for the snapshot.

David, looking in sanlock log - we don't see any error matching this failure,
but the domain 900b1853-e192-4661-a0f9-7c7c396f6f49 has renewal errors.

I guess because sanlock_init_resource is implemented in the library,
not going trough sanlock deamon?

2017-02-20 14:30:09+0100 1050804 [11738]: 900b1853 aio timeout RD
0x7f41d00008c0:0x7f41d00008d0:0x7f41e2afa000 ioto 10 to_count 1
2017-02-20 14:30:09+0100 1050804 [11738]: s3 delta_renew read timeout
10 sec offset 0 /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/ids
2017-02-20 14:30:09+0100 1050804 [11738]: s3 renewal error -202
delta_length 10 last_success 1050773
2017-02-20 14:30:11+0100 1050806 [11738]: 900b1853 aio collect RD
0x7f41d00008c0:0x7f41d00008d0:0x7f41e2afa000 result 1048576:0 match
reap
2017-02-20 14:35:58+0100 1051153 [11738]: 900b1853 aio timeout RD
0x7f41d00008c0:0x7f41d00008d0:0x7f41e2afa000 ioto 10 to_count 2
2017-02-20 14:35:58+0100 1051153 [11738]: s3 delta_renew read timeout
10 sec offset 0 /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/ids
2017-02-20 14:35:58+0100 1051153 [11738]: s3 renewal error -202
delta_length 10 last_success 1051122
2017-02-20 14:36:01+0100 1051156 [11738]: 900b1853 aio collect RD
0x7f41d00008c0:0x7f41d00008d0:0x7f41e2afa000 result 1048576:0 match
reap
2017-02-20 14:44:36+0100 1051671 [11738]: 900b1853 aio timeout RD
0x7f41d00008c0:0x7f41d00008d0:0x7f41e2afa000 ioto 10 to_count 3
2017-02-20 14:44:36+0100 1051671 [11738]: s3 delta_renew read timeout
10 sec offset 0 /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/ids
2017-02-20 14:44:36+0100 1051671 [11738]: s3 renewal error -202
delta_length 10 last_success 1051641
2017-02-20 14:44:37+0100 1051672 [11738]: 900b1853 aio collect RD
0x7f41d00008c0:0x7f41d00008d0:0x7f41e2afa000 result 1048576:0 match
reap
2017-02-20 14:48:02+0100 1051877 [11738]: 900b1853 aio timeout RD
0x7f41d00008c0:0x7f41d00008d0:0x7f41e2afa000 ioto 10 to_count 4
2017-02-20 14:48:02+0100 1051877 [11738]: s3 delta_renew read timeout
10 sec offset 0 /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/ids
2017-02-20 14:48:02+0100 1051877 [11738]: s3 renewal error -202
delta_length 10 last_success 1051846
2017-02-20 14:48:02+0100 1051877 [11738]: 900b1853 aio collect RD
0x7f41d00008c0:0x7f41d00008d0:0x7f41e2afa000 result 1048576:0 match
reap

Gianluca, what is domain 900b1853-e192-4661-a0f9-7c7c396f6f49?

is this the domain you are migrating to in the same time?

Can you share the output of:

    sanlock client renewal -s 900b1853-e192-4661-a0f9-7c7c396f6f49

>>> 2017-02-20 14:41:44,900 ERROR (tasks/8) [storage.TaskManager.Task]
>>> (Task='d694b892-b078-4d86-a035-427ee4fb3b13') Unexpected error (task:870)
>>> Traceback (most recent call last):
>>>   File "/usr/share/vdsm/storage/task.py", line 877, in _run
>>>     return fn(*args, **kargs)
>>>   File "/usr/share/vdsm/storage/task.py", line 333, in run
>>>     return self.cmd(*self.argslist, **self.argsdict)
>>>   File "/usr/lib/python2.7/site-packages/vdsm/storage/securable.py", line
>>> 79, in wrapper
>>>     return method(self, *args, **kwargs)
>>>   File "/usr/share/vdsm/storage/sp.py", line 1929, in createVolume
>>>     initialSize=initialSize)
>>>   File "/usr/share/vdsm/storage/sd.py", line 762, in createVolume
>>>     initialSize=initialSize)
>>>   File "/usr/share/vdsm/storage/volume.py", line 1089, in create
>>>     (volUUID, e))
>>> VolumeCreationError: Error creating a new volume: (u"Volume creation
>>> d0d938bd-1479-49cb-93fb-85b6a32d6cb4 failed: (-202, 'Sanlock resource init
>>> failure', 'Sanlock exception')",)
>>> 2017-02-20 14:41:44,941 INFO  (tasks/8) [storage.Volume] Metadata rollback
>>> for sdUUID=900b1853-e192-4661-a0f9-7c7c396f6f49 offs=8 (blockVolume:448)
>>>
>>>
>>> Was the error generated due to the other migration still in progress?
>>> Is there a limit of concurrent migrations from/to a particular storage
>>> domain?
>>
>> No, maybe your network was overloaded by the concurrent migrations?
>>
>>>
>>> Now I would like to retry, but I see that the disk is in state locked with
>>> hourglass.
>>> The autogenerated snapshot of the failed action was apparently removed with
>>> success as I don't see it.
>>>
>>> How can I proceed to move the disk?
>>>
>>> Thanks in advance,
>>> Gianluca
>>>
>>> _______________________________________________
>>> Users mailing list
>>> Users@ovirt.org
>>> http://lists.ovirt.org/mailman/listinfo/users
>>>
_______________________________________________
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users

Reply via email to