The issue is present in the logs:
2018-05-17 11:50:44,822+01 INFO
[org.ovirt.engine.core.bll.storage.disk.image.VdsmImagePoller]
(DefaultQuartzScheduler1) [39755bb7-9082-40d6-ae5e-64b5b2b5f98e] Command
CopyData id: '84a49b25-0e37-4338-834e-08bd67c42860': the volume lease is
not FREE - the job is running
I tried setting the log level to debug but it seems I have not a
vdsm-client command. All I have is a vdsm-tool command. Is it
equivalent?
Thanks
El 2018-05-17 11:49, Benny Zlotnik escribió:
> By the way, please verify it's the same issue, you should see "the
> volume lease is not FREE - the job is running" in the engine log
>
> On Thu, May 17, 2018 at 1:21 PM, Benny Zlotnik <bzlotnik(a)redhat.com>
> wrote:
>
>> I see because I am on debug level, you need to enable it in order to
>> see
>>
>>
https://www.ovirt.org/develop/developer-guide/vdsm/log-files/ [3]
>>
>> On Thu, 17 May 2018, 13:10 , <nicolas(a)devels.es> wrote:
>>
>>> Hi,
>>>
>>> Thanks. I've checked vdsm logs on all my hosts but the only entry
>>> I can
>>> find grepping by Volume.getInfo is like this:
>>>
>>> 2018-05-17 10:14:54,892+0100 INFO (jsonrpc/0)
>>> [jsonrpc.JsonRpcServer]
>>> RPC call Volume.getInfo succeeded in 0.30 seconds (__init__:539)
>>>
>>> I cannot find a line like yours... any other way on how to obtain
>>> those
>>> parameters. This is an iSCSI based storage FWIW (both source and
>>> destination of the movement).
>>>
>>> Thanks.
>>>
>>> El 2018-05-17 10:01, Benny Zlotnik escribió:
>>>> In the vdsm log you will find the volumeInfo log which looks
>>> like
>>>> this:
>>>>
>>>> 2018-05-17 11:55:03,257+0300 DEBUG (jsonrpc/6)
>>> [jsonrpc.JsonRpcServer]
>>>> Return 'Volume.getInfo' in bridge with {'status':
'OK',
>>> 'domain':
>>>> '5c4d2216-
>>>> 2eb3-4e24-b254-d5f83fde4dbe', 'voltype': 'INTERNAL',
>>> 'description':
>>>>
'{"DiskAlias":"vm_Disk1","DiskDescription":""}',
'parent':
>>>> '00000000-0000-0000-
>>>> 0000-000000000000', 'format': 'RAW',
'generation': 3, 'image':
>>>> 'b8eb8c82-fddd-4fbc-b80d-6ee04c1255bc', 'ctime':
'1526543244',
>>>> 'disktype': 'DATA', '
>>>> legality': 'LEGAL', 'mtime': '0',
'apparentsize': '1073741824',
>>>> 'children': [], 'pool': '', 'capacity':
'1073741824', 'uuid':
>>>> u'7190913d-320c-4fc9-
>>>> a5b3-c55b26aa30f4', 'truesize': '0', 'type':
'SPARSE', 'lease':
>>>> {'path':
>>>>
>>>
>>
> u'/rhev/data-center/mnt/10.35.0.233:_root_storage__domains_sd1/5c4d2216-2e
>>>>
>>>
>>
>
b3-4e24-b254-d5f83fde4dbe/images/b8eb8c82-fddd-4fbc-b80d-6ee04c1255bc/7190913d-320c-4fc9-a5b3-c55b26aa30f4.lease',
>>>> 'owners': [1], 'version': 8L, 'o
>>>> ffset': 0}} (__init__:355)
>>>>
>>>> The lease path in my case is:
>>>> /rhev/data-center/mnt/10.35.0.
>>>
>>
>
[1]233:_root_storage__domains_sd1/5c4d2216-2eb3-4e24-b254-d5f83fde4dbe/images/b8eb8c82-fddd-4fbc-b80d-6ee04c1255bc/7190913d-320c-4fc9-a5b3-c55b26aa30f4.lease
>>>>
>>>> Then you can look in /var/log/sanlock.log
>>>>
>>>> 2018-05-17 11:35:18 243132 [14847]: s2:r9 resource
>>>>
>>>
>>
>
5c4d2216-2eb3-4e24-b254-d5f83fde4dbe:7190913d-320c-4fc9-a5b3-c55b26aa30f4:/rhev/data-center/mnt/10.35.0.233:_root_storage__domains_sd1/5c4d2216-2eb3-4e24-b254-d5f83fde4dbe/images/b8eb8c82-fddd-4fbc-b80d-6ee04c1255bc/7190913d-320c-4fc9-a5b3-c55b26aa30f4.lease:0
>>>> for 2,9,5049
>>>>
>>>> Then you can use this command to unlock, the pid in this case
>>> is 5049
>>>>
>>>> sanlock client release -r RESOURCE -p pid
>>>>
>>>> On Thu, May 17, 2018 at 11:52 AM, Benny Zlotnik
>>> <bzlotnik(a)redhat.com>
>>>> wrote:
>>>>
>>>>> I believe you've hit this
>>>>>
bug: https://bugzilla.redhat.com/show_bug.cgi?id=1565040 [2]
>>> [1]
>>>>>
>>>>> You can try to release the lease manually using the
>>> sanlock client
>>>>> command (there's an example in the comments on the bug),
>>>>> once the lease is free the job will fail and the disk can be
>>> unlock
>>>>>
>>>>> On Thu, May 17, 2018 at 11:05 AM, <nicolas(a)devels.es> wrote:
>>>>>
>>>>>> Hi,
>>>>>>
>>>>>> We're running oVirt 4.1.9 (I know it's not the
recommended
>>>>>> version, but we can't upgrade yet) and recently we had an
>>> issue
>>>>>> with a Storage Domain while a VM was moving a disk. The
>>> Storage
>>>>>> Domain went down for a few minutes, then it got back.
>>>>>>
>>>>>> However, the disk's state has stuck in a 'Migrating:
10%'
>>> state
>>>>>> (see ss-2.png).
>>>>>>
>>>>>> I run the 'unlock_entity.sh' script to try to unlock the
>>> disk,
>>>>>> with these parameters:
>>>>>>
>>>>>> # PGPASSWORD=...
>>>>>> /usr/share/ovirt-engine/setup/dbutils/unlock_entity.sh -t
>>> disk -u
>>>>>> engine -v b4013aba-a936-4a54-bb14-670d3a8b7c38
>>>>>>
>>>>>> The disk's state changed to 'OK', but the actual
state still
>>>>>> states it's migrating (see ss-1.png).
>>>>>>
>>>>>> Calling the script with -t all doesn't make a difference
>>> either.
>>>>>>
>>>>>> Currently, the disk is unmanageable: cannot be deactivated,
>>> moved
>>>>>> or copied, as it says there's a copying operation running
>>> already.
>>>>>>
>>>>>> Could someone provide a way to unlock this disk? I don't
mind
>>>>>> modifying a value directly into the database, I just need the
>>>>>> copying process cancelled.
>>>>>>
>>>>>> Thanks.
>>>>>> _______________________________________________
>>>>>> Users mailing list -- users(a)ovirt.org
>>>>>> To unsubscribe send an email to users-leave(a)ovirt.org
>>>>
>>>>
>>>>
>>>> Links:
>>>> ------
>>>> [1]
https://bugzilla.redhat.com/show_bug.cgi?id=1565040 [2]
>
>
>
> Links:
> ------
> [1]
http://10.35.0.
> [2]
https://bugzilla.redhat.com/show_bug.cgi?id=1565040
> [3]
https://www.ovirt.org/develop/developer-guide/vdsm/log-files/