Which vdsm version are you using?

You can try looking for the image uuid in /var/log/sanlock.log

On Thu, May 17, 2018 at 2:40 PM, <nico...@devels.es> wrote:

> Thanks.
>
> I've been able to see the line in the log, however, the format differs
> slightly from yours.
>
>   2018-05-17 12:24:44,132+0100 DEBUG (jsonrpc/6) [jsonrpc.JsonRpcServer]
> Calling 'Volume.getInfo' in bridge with {u'storagepoolID':
> u'75bf8f48-970f-42bc-8596-f8ab6efb2b63', u'imageID':
> u'b4013aba-a936-4a54-bb14-670d3a8b7c38', u'volumeID':
> u'c2cfbb02-9981-4fb7-baea-7257a824145c', u'storagedomainID':
> u'1876ab86-216f-4a37-a36b-2b5d99fcaad0'} (__init__:556)
> 2018-05-17 12:24:44,689+0100 DEBUG (jsonrpc/6) [jsonrpc.JsonRpcServer]
> Return 'Volume.getInfo' in bridge with {'status': 'OK', 'domain':
> '1876ab86-216f-4a37-a36b-2b5d99fcaad0', 'voltype': 'INTERNAL',
> 'description': 'None', 'parent': 'ea9a0182-329f-4b8f-abe3-e894de95dac0',
> 'format': 'COW', 'generation': 1, 'image': 
> 'b4013aba-a936-4a54-bb14-670d3a8b7c38',
> 'ctime': '1526470759', 'disktype': '2', 'legality': 'LEGAL', 'mtime': '0',
> 'apparentsize': '1073741824', 'children': [], 'pool': '', 'capacity':
> '21474836480', 'uuid': u'c2cfbb02-9981-4fb7-baea-7257a824145c',
> 'truesize': '1073741824', 'type': 'SPARSE', 'lease': {'owners': [8],
> 'version': 1L}} (__init__:582)
>
> As you can see, there's no path field there.
>
> How should I procceed?
>
> El 2018-05-17 12:01, Benny Zlotnik escribió:
>
>> vdsm-client replaces vdsClient, take a look
>> here: https://lists.ovirt.org/pipermail/devel/2016-July/013535.html
>> [4]
>>
>> On Thu, May 17, 2018 at 1:57 PM, <nico...@devels.es> wrote:
>>
>> The issue is present in the logs:
>>>
>>>   2018-05-17 11:50:44,822+01 INFO
>>> [org.ovirt.engine.core.bll.storage.disk.image.VdsmImagePoller]
>>> (DefaultQuartzScheduler1) [39755bb7-9082-40d6-ae5e-64b5b2b5f98e]
>>> Command CopyData id: '84a49b25-0e37-4338-834e-08bd67c42860': the
>>> volume lease is not FREE - the job is running
>>>
>>> I tried setting the log level to debug but it seems I have not a
>>> vdsm-client command. All I have is a vdsm-tool command. Is it
>>> equivalent?
>>>
>>> Thanks
>>>
>>> El 2018-05-17 11:49, Benny Zlotnik escribió:
>>> By the way, please verify it's the same issue, you should see "the
>>> volume lease is not FREE - the job is running" in the engine log
>>>
>>> On Thu, May 17, 2018 at 1:21 PM, Benny Zlotnik
>>> <bzlot...@redhat.com>
>>> wrote:
>>>
>>> I see because I am on debug level, you need to enable it in order
>>> to
>>> see
>>>
>>> https://www.ovirt.org/develop/developer-guide/vdsm/log-files/ [1]
>>>
>>> [3]
>>>
>>> On Thu, 17 May 2018, 13:10 , <nico...@devels.es> wrote:
>>>
>>> Hi,
>>>
>>> Thanks. I've checked vdsm logs on all my hosts but the only entry
>>> I can
>>> find grepping by Volume.getInfo is like this:
>>>
>>>    2018-05-17 10:14:54,892+0100 INFO  (jsonrpc/0)
>>> [jsonrpc.JsonRpcServer]
>>> RPC call Volume.getInfo succeeded in 0.30 seconds (__init__:539)
>>>
>>> I cannot find a line like yours... any other way on how to obtain
>>> those
>>> parameters. This is an iSCSI based storage FWIW (both source and
>>> destination of the movement).
>>>
>>> Thanks.
>>>
>>> El 2018-05-17 10:01, Benny Zlotnik escribió:
>>> In the vdsm log you will find the volumeInfo log which looks
>>> like
>>> this:
>>>
>>> 2018-05-17 11:55:03,257+0300 DEBUG (jsonrpc/6)
>>> [jsonrpc.JsonRpcServer]
>>> Return 'Volume.getInfo' in bridge with {'status': 'OK',
>>> 'domain':
>>> '5c4d2216-
>>> 2eb3-4e24-b254-d5f83fde4dbe', 'voltype': 'INTERNAL',
>>> 'description':
>>> '{"DiskAlias":"vm_Disk1","DiskDescription":""}', 'parent':
>>> '00000000-0000-0000-
>>> 0000-000000000000', 'format': 'RAW', 'generation': 3, 'image':
>>> 'b8eb8c82-fddd-4fbc-b80d-6ee04c1255bc', 'ctime': '1526543244',
>>> 'disktype': 'DATA', '
>>> legality': 'LEGAL', 'mtime': '0', 'apparentsize': '1073741824',
>>> 'children': [], 'pool': '', 'capacity': '1073741824', 'uuid':
>>> u'7190913d-320c-4fc9-
>>> a5b3-c55b26aa30f4', 'truesize': '0', 'type': 'SPARSE', 'lease':
>>> {'path':
>>>
>>
>>  u'/rhev/data-center/mnt/10.35.0.233:_root_storage__
>> domains_sd1/5c4d2216-2e
>>
>>
>>>
>>
>> b3-4e24-b254-d5f83fde4dbe/images/b8eb8c82-fddd-4fbc-b80d-
>> 6ee04c1255bc/7190913d-320c-4fc9-a5b3-c55b26aa30f4.lease',
>>
>> 'owners': [1], 'version': 8L, 'o
>>> ffset': 0}} (__init__:355)
>>>
>>> The lease path in my case is:
>>> /rhev/data-center/mnt/10.35.0. [2]
>>>
>>
>>
>> [1]233:_root_storage__domains_sd1/5c4d2216-2eb3-4e24-b254-d5
>> f83fde4dbe/images/b8eb8c82-fddd-4fbc-b80d-6ee04c1255bc/71909
>> 13d-320c-4fc9-a5b3-c55b26aa30f4.lease
>>
>> Then you can look in /var/log/sanlock.log
>>>
>>> 2018-05-17 11:35:18 243132 [14847]: s2:r9 resource
>>>
>>
>>
>> 5c4d2216-2eb3-4e24-b254-d5f83fde4dbe:7190913d-320c-4fc9-
>> a5b3-c55b26aa30f4:/rhev/data-center/mnt/10.35.0.233:_root_
>> storage__domains_sd1/5c4d2216-2eb3-4e24-b254-d5f83fde4dbe/
>> images/b8eb8c82-fddd-4fbc-b80d-6ee04c1255bc/7190913d-
>> 320c-4fc9-a5b3-c55b26aa30f4.lease:0
>>
>> for 2,9,5049
>>>
>>> Then you can use this command to unlock, the pid in this case
>>> is 5049
>>>
>>> sanlock client release -r RESOURCE -p pid
>>>
>>> On Thu, May 17, 2018 at 11:52 AM, Benny Zlotnik
>>> <bzlot...@redhat.com>
>>> wrote:
>>>
>>> I believe you've hit this
>>> bug: https://bugzilla.redhat.com/show_bug.cgi?id=1565040 [3] [2]
>>>
>>
>>  [1]
>>
>> You can try to release the lease manually using the
>>>>
>>>  sanlock client
>>
>> command (there's an example in the comments on the bug),
>>>> once the lease is free the job will fail and the disk can be
>>>>
>>>  unlock
>>
>> On Thu, May 17, 2018 at 11:05 AM, <nico...@devels.es> wrote:
>>>
>>> Hi,
>>>
>>> We're running oVirt 4.1.9 (I know it's not the recommended
>>> version, but we can't upgrade yet) and recently we had an
>>>
>>  issue
>>
>> with a Storage Domain while a VM was moving a disk. The
>>>
>>  Storage
>>
>> Domain went down for a few minutes, then it got back.
>>>
>>> However, the disk's state has stuck in a 'Migrating: 10%'
>>>
>>  state
>>
>> (see ss-2.png).
>>>
>>> I run the 'unlock_entity.sh' script to try to unlock the
>>>
>>  disk,
>>
>> with these parameters:
>>>
>>>  # PGPASSWORD=...
>>> /usr/share/ovirt-engine/setup/dbutils/unlock_entity.sh -t
>>>
>>  disk -u
>>
>> engine -v b4013aba-a936-4a54-bb14-670d3a8b7c38
>>>
>>> The disk's state changed to 'OK', but the actual state still
>>> states it's migrating (see ss-1.png).
>>>
>>> Calling the script with -t all doesn't make a difference
>>>
>>  either.
>>
>> Currently, the disk is unmanageable: cannot be deactivated,
>>>
>>  moved
>>
>> or copied, as it says there's a copying operation running
>>>
>>  already.
>>
>> Could someone provide a way to unlock this disk? I don't mind
>>> modifying a value directly into the database, I just need the
>>> copying process cancelled.
>>>
>>> Thanks.
>>> _______________________________________________
>>> Users mailing list -- users@ovirt.org
>>> To unsubscribe send an email to users-le...@ovirt.org
>>>
>>
>>  Links:
>>  ------
>>  [1] https://bugzilla.redhat.com/show_bug.cgi?id=1565040 [3] [2]
>>
>>  Links:
>>  ------
>>  [1] http://10.35.0 [5].
>>  [2] https://bugzilla.redhat.com/show_bug.cgi?id=1565040 [3]
>>  [3] https://www.ovirt.org/develop/developer-guide/vdsm/log-files/ [1]
>>
>>
>>
>> Links:
>> ------
>> [1] https://www.ovirt.org/develop/developer-guide/vdsm/log-files/
>> [2] http://10.35.0.
>> [3] https://bugzilla.redhat.com/show_bug.cgi?id=1565040
>> [4] https://lists.ovirt.org/pipermail/devel/2016-July/013535.html
>> [5] http://10.35.0
>>
> _______________________________________________
> Users mailing list -- users@ovirt.org
> To unsubscribe send an email to users-le...@ovirt.org
>
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org

Reply via email to