[ovirt-users] oVirt gluster sanlock issue

Abi Askushi rightkicktech at gmail.com
Tue Jun 6 12:24:31 UTC 2017


Also I see this article from Redhat, that mentions if 4K sectors are
supported, but I am not able to read it as I don't have a subscription:

https://access.redhat.com/solutions/56494

Its hard to believe that 4K drives have not been used from others on oVirt
deployments.

Alex


On Tue, Jun 6, 2017 at 3:18 PM, Abi Askushi <rightkicktech at gmail.com> wrote:

> Hi Krutika,
>
> My comments inline.
>
> Also attached the strace of:
> strace -y -ff -o /root/512-trace-on-root.log dd if=/dev/zero
> of=/mnt/test2.img oflag=direct bs=512 count=1
>
> and of:
> strace -y -ff -o /root/4096-trace-on-root.log dd if=/dev/zero
> of=/mnt/test2.img oflag=direct bs=4096 count=16
>
> I have mounted gluster volume at /mnt.
> The dd with bs=4096 is successful.
>
> The gluster mount log gives only the following:
> [2017-06-06 12:04:54.102576] W [MSGID: 114031] [client-rpc-fops.c:854:client3_3_writev_cbk]
> 0-engine-client-0: remote operation failed [Invalid argument]
> [2017-06-06 12:04:54.102591] W [MSGID: 114031] [client-rpc-fops.c:854:client3_3_writev_cbk]
> 0-engine-client-1: remote operation failed [Invalid argument]
> [2017-06-06 12:04:54.103355] W [fuse-bridge.c:2312:fuse_writev_cbk]
> 0-glusterfs-fuse: 205: WRITE => -1 gfid=075ab3a5-0274-4f07-a075-2748c3b4d394
> fd=0x7faf1d08706c (Transport endpoint is not connected)
>
> The gluster brick log gives:
> [2017-06-06 12:07:03.793080] E [MSGID: 113072] [posix.c:3453:posix_writev]
> 0-engine-posix: write failed: offset 0, [Invalid argument]
> [2017-06-06 12:07:03.793172] E [MSGID: 115067] [server-rpc-fops.c:1346:server_writev_cbk]
> 0-engine-server: 291: WRITEV 0 (075ab3a5-0274-4f07-a075-2748c3b4d394) ==>
> (Invalid argument) [Invalid argument]
>
>
>
> On Tue, Jun 6, 2017 at 12:50 PM, Krutika Dhananjay <kdhananj at redhat.com>
> wrote:
>
>> OK.
>>
>> So for the 'Transport endpoint is not connected' issue, could you share
>> the mount and brick logs?
>>
>> Hmmm.. 'Invalid argument' error even on the root partition. What if you
>> change bs to 4096 and run?
>>
> If I use bs=4096 the dd is successful on /root and at gluster mounted
> volume.
>
>>
>> The logs I showed in my earlier mail shows that gluster is merely
>> returning the error it got from the disk file system where the
>> brick is hosted. But you're right about the fact that the offset 127488
>> is not 4K-aligned.
>>
>> If the dd on /root worked for you with bs=4096, could you try the same
>> directly on gluster mount point on a dummy file and capture the strace
>> output of dd?
>> You can perhaps reuse your existing gluster volume by mounting it at
>> another location and doing the dd.
>> Here's what you need to execute:
>>
>> strace -ff -T -p <pid-of-mount-process> -o <path-to-the-file-where-you-want-the-output-saved>`
>>
>> FWIW, here's something I found in man(2) open:
>>
>>
>>
>>
>> *Under  Linux  2.4,  transfer  sizes,  and  the alignment of the user
>> buffer and the file offset must all be multiples of the logical block size
>> of the filesystem.  Since Linux 2.6.0, alignment to the logical block size
>> of the       underlying storage (typically 512 bytes) suffices.  The
>> logical block size can be determined using the ioctl(2) BLKSSZGET operation
>> or from the shell using the command:           blockdev --getss*
>>
> Please note also that the physical disks are of 4K sector size (native).
> Thus OS is having 4096/4096 local/physical sector size.
> [root at v0 ~]# blockdev --getss /dev/sda
> 4096
> [root at v0 ~]# blockdev --getpbsz /dev/sda
> 4096
>
>>
>>
>> -Krutika
>>
>>
>> On Tue, Jun 6, 2017 at 1:18 AM, Abi Askushi <rightkicktech at gmail.com>
>> wrote:
>>
>>> Also when testing with dd i get the following:
>>>
>>> *Testing on the gluster mount: *
>>> dd if=/dev/zero of=/rhev/data-center/mnt/glusterSD/10.100.100.1:
>>> _engine/test2.img oflag=direct bs=512 count=1
>>> dd: error writing β/rhev/data-center/mnt/glusterSD/10.100.100.1:
>>> _engine/test2.imgβ: *Transport endpoint is not connected*
>>> 1+0 records in
>>> 0+0 records out
>>> 0 bytes (0 B) copied, 0.00336755 s, 0.0 kB/s
>>>
>>> *Testing on the /root directory (XFS): *
>>> dd if=/dev/zero of=/test2.img oflag=direct bs=512 count=1
>>> dd: error writing β/test2.imgβ:* Invalid argument*
>>> 1+0 records in
>>> 0+0 records out
>>> 0 bytes (0 B) copied, 0.000321239 s, 0.0 kB/s
>>>
>>> Seems that the gluster is trying to do the same and fails.
>>>
>>>
>>>
>>> On Mon, Jun 5, 2017 at 10:10 PM, Abi Askushi <rightkicktech at gmail.com>
>>> wrote:
>>>
>>>> The question that rises is what is needed to make gluster aware of the
>>>> 4K physical sectors presented to it (the logical sector is also 4K). The
>>>> offset (127488) at the log does not seem aligned at 4K.
>>>>
>>>> Alex
>>>>
>>>> On Mon, Jun 5, 2017 at 2:47 PM, Abi Askushi <rightkicktech at gmail.com>
>>>> wrote:
>>>>
>>>>> Hi Krutika,
>>>>>
>>>>> I am saying that I am facing this issue with 4k drives. I never
>>>>> encountered this issue with 512 drives.
>>>>>
>>>>> Alex
>>>>>
>>>>> On Jun 5, 2017 14:26, "Krutika Dhananjay" <kdhananj at redhat.com> wrote:
>>>>>
>>>>>> This seems like a case of O_DIRECT reads and writes gone wrong,
>>>>>> judging by the 'Invalid argument' errors.
>>>>>>
>>>>>> The two operations that have failed on gluster bricks are:
>>>>>>
>>>>>> [2017-06-05 09:40:39.428979] E [MSGID: 113072]
>>>>>> [posix.c:3453:posix_writev] 0-engine-posix: write failed: offset 0,
>>>>>> [Invalid argument]
>>>>>> [2017-06-05 09:41:00.865760] E [MSGID: 113040]
>>>>>> [posix.c:3178:posix_readv] 0-engine-posix: read failed on
>>>>>> gfid=8c94f658-ac3c-4e3a-b368-8c038513a914, fd=0x7f408584c06c,
>>>>>> offset=127488 size=512, buf=0x7f4083c0b000 [Invalid argument]
>>>>>>
>>>>>> But then, both the write and the read have 512byte-aligned offset,
>>>>>> size and buf address (which is correct).
>>>>>>
>>>>>> Are you saying you don't see this issue with 4K block-size?
>>>>>>
>>>>>> -Krutika
>>>>>>
>>>>>> On Mon, Jun 5, 2017 at 3:21 PM, Abi Askushi <rightkicktech at gmail.com>
>>>>>> wrote:
>>>>>>
>>>>>>> Hi Sahina,
>>>>>>>
>>>>>>> Attached are the logs. Let me know if sth else is needed.
>>>>>>>
>>>>>>> I have 5 disks (with 4K physical sector) in RAID5. The RAID has 64K
>>>>>>> stripe size at the moment.
>>>>>>> I have prepared the storage as below:
>>>>>>>
>>>>>>> pvcreate --dataalignment 256K /dev/sda4
>>>>>>> vgcreate --physicalextentsize 256K gluster /dev/sda4
>>>>>>>
>>>>>>> lvcreate -n engine --size 120G gluster
>>>>>>> mkfs.xfs -f -i size=512 /dev/gluster/engine
>>>>>>>
>>>>>>> Thanx,
>>>>>>> Alex
>>>>>>>
>>>>>>> On Mon, Jun 5, 2017 at 12:14 PM, Sahina Bose <sabose at redhat.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> Can we have the gluster mount logs and brick logs to check if it's
>>>>>>>> the same issue?
>>>>>>>>
>>>>>>>> On Sun, Jun 4, 2017 at 11:21 PM, Abi Askushi <
>>>>>>>> rightkicktech at gmail.com> wrote:
>>>>>>>>
>>>>>>>>> I clean installed everything and ran into the same.
>>>>>>>>> I then ran gdeploy and encountered the same issue when deploying
>>>>>>>>> engine.
>>>>>>>>> Seems that gluster (?) doesn't like 4K sector drives. I am not
>>>>>>>>> sure if it has to do with alignment. The weird thing is that gluster
>>>>>>>>> volumes are all ok, replicating normally and no split brain is reported.
>>>>>>>>>
>>>>>>>>> The solution to the mentioned bug (1386443
>>>>>>>>> <https://bugzilla.redhat.com/show_bug.cgi?id=1386443>) was to
>>>>>>>>> format with 512 sector size, which for my case is not an option:
>>>>>>>>>
>>>>>>>>> mkfs.xfs -f -i size=512 -s size=512 /dev/gluster/engine
>>>>>>>>> illegal sector size 512; hw sector is 4096
>>>>>>>>>
>>>>>>>>> Is there any workaround to address this?
>>>>>>>>>
>>>>>>>>> Thanx,
>>>>>>>>> Alex
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On Sun, Jun 4, 2017 at 5:48 PM, Abi Askushi <
>>>>>>>>> rightkicktech at gmail.com> wrote:
>>>>>>>>>
>>>>>>>>>> Hi Maor,
>>>>>>>>>>
>>>>>>>>>> My disk are of 4K block size and from this bug seems that gluster
>>>>>>>>>> replica needs 512B block size.
>>>>>>>>>> Is there a way to make gluster function with 4K drives?
>>>>>>>>>>
>>>>>>>>>> Thank you!
>>>>>>>>>>
>>>>>>>>>> On Sun, Jun 4, 2017 at 2:34 PM, Maor Lipchuk <mlipchuk at redhat.com
>>>>>>>>>> > wrote:
>>>>>>>>>>
>>>>>>>>>>> Hi Alex,
>>>>>>>>>>>
>>>>>>>>>>> I saw a bug that might be related to the issue you encountered at
>>>>>>>>>>> https://bugzilla.redhat.com/show_bug.cgi?id=1386443
>>>>>>>>>>>
>>>>>>>>>>> Sahina, maybe you have any advise? Do you think that BZ1386443is
>>>>>>>>>>> related?
>>>>>>>>>>>
>>>>>>>>>>> Regards,
>>>>>>>>>>> Maor
>>>>>>>>>>>
>>>>>>>>>>> On Sat, Jun 3, 2017 at 8:45 PM, Abi Askushi <
>>>>>>>>>>> rightkicktech at gmail.com> wrote:
>>>>>>>>>>> > Hi All,
>>>>>>>>>>> >
>>>>>>>>>>> > I have installed successfully several times oVirt (version
>>>>>>>>>>> 4.1) with 3 nodes
>>>>>>>>>>> > on top glusterfs.
>>>>>>>>>>> >
>>>>>>>>>>> > This time, when trying to configure the same setup, I am
>>>>>>>>>>> facing the
>>>>>>>>>>> > following issue which doesn't seem to go away. During
>>>>>>>>>>> installation i get the
>>>>>>>>>>> > error:
>>>>>>>>>>> >
>>>>>>>>>>> > Failed to execute stage 'Misc configuration': Cannot acquire
>>>>>>>>>>> host id:
>>>>>>>>>>> > (u'a5a6b0e7-fc3f-4838-8e26-c8b4d5e5e922',
>>>>>>>>>>> SanlockException(22, 'Sanlock
>>>>>>>>>>> > lockspace add failure', 'Invalid argument'))
>>>>>>>>>>> >
>>>>>>>>>>> > The only different in this setup is that instead of standard
>>>>>>>>>>> partitioning i
>>>>>>>>>>> > have GPT partitioning and the disks have 4K block size instead
>>>>>>>>>>> of 512.
>>>>>>>>>>> >
>>>>>>>>>>> > The /var/log/sanlock.log has the following lines:
>>>>>>>>>>> >
>>>>>>>>>>> > 2017-06-03 19:21:15+0200 23450 [943]: s9 lockspace
>>>>>>>>>>> > ba6bd862-c2b8-46e7-b2c8-91e4a5bb2047:250:/rhev/data-center/m
>>>>>>>>>>> nt/_var_lib_ovirt-hosted-engin-setup_tmptjkIDI/ba6bd862-c2b8
>>>>>>>>>>> -46e7-b2c8-91e4a5bb2047/dom_md/ids:0
>>>>>>>>>>> > 2017-06-03 19:21:36+0200 23471 [944]: s9:r5 resource
>>>>>>>>>>> > ba6bd862-c2b8-46e7-b2c8-91e4a5bb2047:SDM:/rhev/data-center/m
>>>>>>>>>>> nt/_var_lib_ovirt-hosted-engine-setup_tmptjkIDI/ba6bd862-c2b
>>>>>>>>>>> 8-46e7-b2c8-91e4a5bb2047/dom_md/leases:1048576
>>>>>>>>>>> > for 2,9,23040
>>>>>>>>>>> > 2017-06-03 19:21:36+0200 23471 [943]: s10 lockspace
>>>>>>>>>>> > a5a6b0e7-fc3f-4838-8e26-c8b4d5e5e922:250:/rhev/data-center/m
>>>>>>>>>>> nt/glusterSD/10.100.100.1:_engine/a5a6b0e7-fc3f-4838-8e26-c8
>>>>>>>>>>> b4d5e5e922/dom_md/ids:0
>>>>>>>>>>> > 2017-06-03 19:21:36+0200 23471 [23522]: a5a6b0e7 aio collect RD
>>>>>>>>>>> > 0x7f59b00008c0:0x7f59b00008d0:0x7f59b0101000 result -22:0
>>>>>>>>>>> match res
>>>>>>>>>>> > 2017-06-03 19:21:36+0200 23471 [23522]: read_sectors
>>>>>>>>>>> delta_leader offset
>>>>>>>>>>> > 127488 rv -22
>>>>>>>>>>> > /rhev/data-center/mnt/glusterSD/10.100.100.1:_engine/a5a6b0e
>>>>>>>>>>> 7-fc3f-4838-8e26-c8b4d5e5e922/dom_md/ids
>>>>>>>>>>> > 2017-06-03 19:21:37+0200 23472 [930]: s9 host 250 1 23450
>>>>>>>>>>> > 88c2244c-a782-40ed-9560-6cfa4d46f853.v0.neptune
>>>>>>>>>>> > 2017-06-03 19:21:37+0200 23472 [943]: s10 add_lockspace fail
>>>>>>>>>>> result -22
>>>>>>>>>>> >
>>>>>>>>>>> > And /var/log/vdsm/vdsm.log says:
>>>>>>>>>>> >
>>>>>>>>>>> > 2017-06-03 19:19:38,176+0200 WARN  (jsonrpc/3)
>>>>>>>>>>> > [storage.StorageServer.MountConnection] Using user specified
>>>>>>>>>>> > backup-volfile-servers option (storageServer:253)
>>>>>>>>>>> > 2017-06-03 19:21:12,379+0200 WARN  (periodic/1) [throttled]
>>>>>>>>>>> MOM not
>>>>>>>>>>> > available. (throttledlog:105)
>>>>>>>>>>> > 2017-06-03 19:21:12,380+0200 WARN  (periodic/1) [throttled]
>>>>>>>>>>> MOM not
>>>>>>>>>>> > available, KSM stats will be missing. (throttledlog:105)
>>>>>>>>>>> > 2017-06-03 19:21:14,714+0200 WARN  (jsonrpc/1)
>>>>>>>>>>> > [storage.StorageServer.MountConnection] Using user specified
>>>>>>>>>>> > backup-volfile-servers option (storageServer:253)
>>>>>>>>>>> > 2017-06-03 19:21:15,515+0200 ERROR (jsonrpc/4)
>>>>>>>>>>> [storage.initSANLock] Cannot
>>>>>>>>>>> > initialize SANLock for domain a5a6b0e7-fc3f-4838-8e26-c8b4d5
>>>>>>>>>>> e5e922
>>>>>>>>>>> > (clusterlock:238)
>>>>>>>>>>> > Traceback (most recent call last):
>>>>>>>>>>> >   File "/usr/lib/python2.7/site-packa
>>>>>>>>>>> ges/vdsm/storage/clusterlock.py", line
>>>>>>>>>>> > 234, in initSANLock
>>>>>>>>>>> >     sanlock.init_lockspace(sdUUID, idsPath)
>>>>>>>>>>> > SanlockException: (107, 'Sanlock lockspace init failure',
>>>>>>>>>>> 'Transport
>>>>>>>>>>> > endpoint is not connected')
>>>>>>>>>>> > 2017-06-03 19:21:15,515+0200 WARN  (jsonrpc/4)
>>>>>>>>>>> > [storage.StorageDomainManifest] lease did not initialize
>>>>>>>>>>> successfully
>>>>>>>>>>> > (sd:557)
>>>>>>>>>>> > Traceback (most recent call last):
>>>>>>>>>>> >   File "/usr/share/vdsm/storage/sd.py", line 552, in
>>>>>>>>>>> initDomainLock
>>>>>>>>>>> >     self._domainLock.initLock(self.getDomainLease())
>>>>>>>>>>> >   File "/usr/lib/python2.7/site-packa
>>>>>>>>>>> ges/vdsm/storage/clusterlock.py", line
>>>>>>>>>>> > 271, in initLock
>>>>>>>>>>> >     initSANLock(self._sdUUID, self._idsPath, lease)
>>>>>>>>>>> >   File "/usr/lib/python2.7/site-packa
>>>>>>>>>>> ges/vdsm/storage/clusterlock.py", line
>>>>>>>>>>> > 239, in initSANLock
>>>>>>>>>>> >     raise se.ClusterLockInitError()
>>>>>>>>>>> > ClusterLockInitError: Could not initialize cluster lock: ()
>>>>>>>>>>> > 2017-06-03 19:21:37,867+0200 ERROR (jsonrpc/2)
>>>>>>>>>>> [storage.StoragePool] Create
>>>>>>>>>>> > pool hosted_datacenter canceled  (sp:655)
>>>>>>>>>>> > Traceback (most recent call last):
>>>>>>>>>>> >   File "/usr/share/vdsm/storage/sp.py", line 652, in create
>>>>>>>>>>> >     self.attachSD(sdUUID)
>>>>>>>>>>> >   File "/usr/lib/python2.7/site-packa
>>>>>>>>>>> ges/vdsm/storage/securable.py", line
>>>>>>>>>>> > 79, in wrapper
>>>>>>>>>>> >     return method(self, *args, **kwargs)
>>>>>>>>>>> >   File "/usr/share/vdsm/storage/sp.py", line 971, in attachSD
>>>>>>>>>>> >     dom.acquireHostId(self.id)
>>>>>>>>>>> >   File "/usr/share/vdsm/storage/sd.py", line 790, in
>>>>>>>>>>> acquireHostId
>>>>>>>>>>> >     self._manifest.acquireHostId(hostId, async)
>>>>>>>>>>> >   File "/usr/share/vdsm/storage/sd.py", line 449, in
>>>>>>>>>>> acquireHostId
>>>>>>>>>>> >     self._domainLock.acquireHostId(hostId, async)
>>>>>>>>>>> >   File "/usr/lib/python2.7/site-packa
>>>>>>>>>>> ges/vdsm/storage/clusterlock.py", line
>>>>>>>>>>> > 297, in acquireHostId
>>>>>>>>>>> >     raise se.AcquireHostIdFailure(self._sdUUID, e)
>>>>>>>>>>> > AcquireHostIdFailure: Cannot acquire host id:
>>>>>>>>>>> > (u'a5a6b0e7-fc3f-4838-8e26-c8b4d5e5e922',
>>>>>>>>>>> SanlockException(22, 'Sanlock
>>>>>>>>>>> > lockspace add failure', 'Invalid argument'))
>>>>>>>>>>> > 2017-06-03 19:21:37,870+0200 ERROR (jsonrpc/2)
>>>>>>>>>>> [storage.StoragePool] Domain
>>>>>>>>>>> > ba6bd862-c2b8-46e7-b2c8-91e4a5bb2047 detach from MSD
>>>>>>>>>>> > ba6bd862-c2b8-46e7-b2c8-91e4a5bb2047 Ver 1 failed. (sp:528)
>>>>>>>>>>> > Traceback (most recent call last):
>>>>>>>>>>> >   File "/usr/share/vdsm/storage/sp.py", line 525, in
>>>>>>>>>>> __cleanupDomains
>>>>>>>>>>> >     self.detachSD(sdUUID)
>>>>>>>>>>> >   File "/usr/lib/python2.7/site-packa
>>>>>>>>>>> ges/vdsm/storage/securable.py", line
>>>>>>>>>>> > 79, in wrapper
>>>>>>>>>>> >     return method(self, *args, **kwargs)
>>>>>>>>>>> >   File "/usr/share/vdsm/storage/sp.py", line 1046, in detachSD
>>>>>>>>>>> >     raise se.CannotDetachMasterStorageDomain(sdUUID)
>>>>>>>>>>> > CannotDetachMasterStorageDomain: Illegal action:
>>>>>>>>>>> > (u'ba6bd862-c2b8-46e7-b2c8-91e4a5bb2047',)
>>>>>>>>>>> > 2017-06-03 19:21:37,872+0200 ERROR (jsonrpc/2)
>>>>>>>>>>> [storage.StoragePool] Domain
>>>>>>>>>>> > a5a6b0e7-fc3f-4838-8e26-c8b4d5e5e922 detach from MSD
>>>>>>>>>>> > ba6bd862-c2b8-46e7-b2c8-91e4a5bb2047 Ver 1 failed. (sp:528)
>>>>>>>>>>> > Traceback (most recent call last):
>>>>>>>>>>> >   File "/usr/share/vdsm/storage/sp.py", line 525, in
>>>>>>>>>>> __cleanupDomains
>>>>>>>>>>> >     self.detachSD(sdUUID)
>>>>>>>>>>> >   File "/usr/lib/python2.7/site-packa
>>>>>>>>>>> ges/vdsm/storage/securable.py", line
>>>>>>>>>>> > 79, in wrapper
>>>>>>>>>>> >     return method(self, *args, **kwargs)
>>>>>>>>>>> >   File "/usr/share/vdsm/storage/sp.py", line 1043, in detachSD
>>>>>>>>>>> >     self.validateAttachedDomain(dom)
>>>>>>>>>>> >   File "/usr/lib/python2.7/site-packa
>>>>>>>>>>> ges/vdsm/storage/securable.py", line
>>>>>>>>>>> > 79, in wrapper
>>>>>>>>>>> >     return method(self, *args, **kwargs)
>>>>>>>>>>> >   File "/usr/share/vdsm/storage/sp.py", line 542, in
>>>>>>>>>>> validateAttachedDomain
>>>>>>>>>>> >     self.validatePoolSD(dom.sdUUID)
>>>>>>>>>>> >   File "/usr/lib/python2.7/site-packa
>>>>>>>>>>> ges/vdsm/storage/securable.py", line
>>>>>>>>>>> > 79, in wrapper
>>>>>>>>>>> >     return method(self, *args, **kwargs)
>>>>>>>>>>> >   File "/usr/share/vdsm/storage/sp.py", line 535, in
>>>>>>>>>>> validatePoolSD
>>>>>>>>>>> >     raise se.StorageDomainNotMemberOfPool(self.spUUID, sdUUID)
>>>>>>>>>>> > StorageDomainNotMemberOfPool: Domain is not member in pool:
>>>>>>>>>>> > u'pool=a1e7e9dd-0cf4-41ae-ba13-36297ed66309,
>>>>>>>>>>> > domain=a5a6b0e7-fc3f-4838-8e26-c8b4d5e5e922'
>>>>>>>>>>> > 2017-06-03 19:21:40,063+0200 ERROR (jsonrpc/2)
>>>>>>>>>>> [storage.TaskManager.Task]
>>>>>>>>>>> > (Task='a2476a33-26f8-4ebd-876d-02fe5d13ef78') Unexpected
>>>>>>>>>>> error (task:870)
>>>>>>>>>>> > Traceback (most recent call last):
>>>>>>>>>>> >  File "/usr/share/vdsm/storage/task.py", line 877, in _run
>>>>>>>>>>> >     return fn(*args, **kargs)
>>>>>>>>>>> >   File "/usr/lib/python2.7/site-packages/vdsm/logUtils.py",
>>>>>>>>>>> line 52, in
>>>>>>>>>>> > wrapper
>>>>>>>>>>> >     res = f(*args, **kwargs)
>>>>>>>>>>> >   File "/usr/share/vdsm/storage/hsm.py", line 959, in
>>>>>>>>>>> createStoragePool
>>>>>>>>>>> >     leaseParams)
>>>>>>>>>>> >   File "/usr/share/vdsm/storage/sp.py", line 652, in create
>>>>>>>>>>> >     self.attachSD(sdUUID)
>>>>>>>>>>> >   File "/usr/lib/python2.7/site-packa
>>>>>>>>>>> ges/vdsm/storage/securable.py", line
>>>>>>>>>>> > 79, in wrapper
>>>>>>>>>>> >     return method(self, *args, **kwargs)
>>>>>>>>>>> >   File "/usr/share/vdsm/storage/sp.py", line 971, in attachSD
>>>>>>>>>>> >     dom.acquireHostId(self.id)
>>>>>>>>>>> >   File "/usr/share/vdsm/storage/sd.py", line 790, in
>>>>>>>>>>> acquireHostId
>>>>>>>>>>> >     self._manifest.acquireHostId(hostId, async)
>>>>>>>>>>> >   File "/usr/share/vdsm/storage/sd.py", line 449, in
>>>>>>>>>>> acquireHostId
>>>>>>>>>>> >     self._domainLock.acquireHostId(hostId, async)
>>>>>>>>>>> >   File "/usr/lib/python2.7/site-packa
>>>>>>>>>>> ges/vdsm/storage/clusterlock.py", line
>>>>>>>>>>> > 297, in acquireHostId
>>>>>>>>>>> >     raise se.AcquireHostIdFailure(self._sdUUID, e)
>>>>>>>>>>> > AcquireHostIdFailure: Cannot acquire host id:
>>>>>>>>>>> > (u'a5a6b0e7-fc3f-4838-8e26-c8b4d5e5e922',
>>>>>>>>>>> SanlockException(22, 'Sanlock
>>>>>>>>>>> > lockspace add failure', 'Invalid argument'))
>>>>>>>>>>> > 2017-06-03 19:21:40,067+0200 ERROR (jsonrpc/2)
>>>>>>>>>>> [storage.Dispatcher]
>>>>>>>>>>> > {'status': {'message': "Cannot acquire host id:
>>>>>>>>>>> > (u'a5a6b0e7-fc3f-4838-8e26-c8b4d5e5e922',
>>>>>>>>>>> SanlockException(22, 'Sanlock
>>>>>>>>>>> > lockspace add failure', 'Invalid argument'))", 'code': 661}}
>>>>>>>>>>> (dispatcher:77)
>>>>>>>>>>> >
>>>>>>>>>>> > The gluster volume prepared for engine storage is online and
>>>>>>>>>>> no split brain
>>>>>>>>>>> > is reported. I don't understand what needs to be done to
>>>>>>>>>>> overcome this. Any
>>>>>>>>>>> > idea will be appreciated.
>>>>>>>>>>> >
>>>>>>>>>>> > Thank you,
>>>>>>>>>>> > Alex
>>>>>>>>>>> >
>>>>>>>>>>> > _______________________________________________
>>>>>>>>>>> > Users mailing list
>>>>>>>>>>> > Users at ovirt.org
>>>>>>>>>>> > http://lists.ovirt.org/mailman/listinfo/users
>>>>>>>>>>> >
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>> _______________________________________________
>>>>>>>>> Users mailing list
>>>>>>>>> Users at ovirt.org
>>>>>>>>> http://lists.ovirt.org/mailman/listinfo/users
>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>> _______________________________________________
>>>>>>> Users mailing list
>>>>>>> Users at ovirt.org
>>>>>>> http://lists.ovirt.org/mailman/listinfo/users
>>>>>>>
>>>>>>>
>>>>>>
>>>>
>>>
>>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20170606/e30a8d7a/attachment-0001.html>


More information about the Users mailing list