Appreciate the updates you've been posting.  It's concerning to me as a Gluster user as well. It would be nice to figure out what happened here. 

On Thu, Feb 6, 2020 at 11:43 AM Christian Reiss <email@christian-reiss.de> wrote:
Hey,

For prosperity: Sadly the only way to fix this was to re-init (wipe)
gluster and start from scratch.

-Chris.

On 03/02/2020 19:23, Strahil Nikolov wrote:
> On February 3, 2020 2:29:55 PM GMT+02:00, Christian Reiss <email@christian-reiss.de> wrote:
>> Ugh,
>>
>> disregarding off all previous stamenets:
>>
>> new findinds: vdsm user can NOT read files larger than 64mb. Root can.
>>
>> [vdsm@node02:/rhev/data-cente[...]c51d8a18370] $ for i in 60 62 64 66
>> 68
>> ; do dd if=/dev/urandom of=file-$i bs=1M count=$i ; done
>>
>> [vdsm@node03:/rhev/data-cente[...]c51d8a18370] $ for i in 60 62 64 66
>> 68
>> ; do echo $i ; dd if=file-$i of=/dev/null ; done
>> 60
>> 122880+0 records in
>> 122880+0 records out
>> 62914560 bytes (63 MB) copied, 0.15656 s, 402 MB/s
>> 62
>> 126976+0 records in
>> 126976+0 records out
>> 65011712 bytes (65 MB) copied, 0.172463 s, 377 MB/s
>> 64
>> 131072+0 records in
>> 131072+0 records out
>> 67108864 bytes (67 MB) copied, 0.180701 s, 371 MB/s
>> 66
>> dd: error reading ‘file-66’: Permission denied
>> 131072+0 records in
>> 131072+0 records out
>> 67108864 bytes (67 MB) copied, 0.105236 s, 638 MB/s
>> 68
>> dd: error reading ‘file-68’: Permission denied
>> 131072+0 records in
>> 131072+0 records out
>> 67108864 bytes (67 MB) copied, 0.17046 s, 394 MB/s
>>
>>
>> The files appeared instantly on all nodes, Writing large files work,
>> however. Writing large files seem to work.
>>
>> I think this is the core issue.
>>
>>
>> On 03/02/2020 12:22, Christian Reiss wrote:
>>> Further findings:
>>>
>>> - modified data gets written to local node, not across gluster.
>>> - vdsm user can create _new_ files on the cluster, this gets synced
>>> immediatly.
>>> - vdsm can modify, across all nodes newly created files, changes
>> apply
>>> immediately.
>>>
>>> I think vdsm user can not modify already existing files over the
>>> gluster. Something selinux?
>>>
>>> -Chris.
>>>
>>> On 03/02/2020 11:46, Christian Reiss wrote:
>>>> Hey,
>>>>
>>>> I think I am barking up the right tree with something (else) here;
>>>> Note the timestamps & id's:
>>>>
>>>>
>>>> dd'ing a disk image as vdsm user, try 1:
>>>>
>>>>
>> [vdsm@node03:/rhev/data-center/mnt/glusterSD/node01.dc-dus.dalason.net:_ssd__storage/fec2eb5e-21b5-496b-9ea5-f718b2cb5556/images/4a55b9c0-d550-4ecb-8dd1-cc1f24f2c7ac]
>>
>>>> $ date ; id ; dd if=5fca6d0e-e320-425b-a89a-f80563461add | pv  | dd
>>>> of=/dev/null
>>>> Mon  3 Feb 11:39:13 CET 2020
>>>> uid=36(vdsm) gid=36(kvm) groups=36(kvm),107(qemu),179(sanlock)
>>>> context=unconfined_u:unconfined_r:unconfined_t:s0-s0:c0.c1023
>>>> dd: error reading ‘5fca6d0e-e320-425b-a89a-f80563461add’: Permission
>>
>>>> denied
>>>> 131072+0 records in
>>>> 131072+0 records out
>>>> 67108864 bytes (67 MB) copied, 0.169465 s, 396 MB/s
>>>>     64MiB 0:00:00 [ 376MiB/s] [  <=>
>>>>        ]
>>>> 131072+0 records in
>>>> 131072+0 records out
>>>> 67108864 bytes (67 MB) copied, 0.171726 s, 391 MB/s
>>>>
>>>>
>>>> try 2, directly afterward:
>>>>
>>>>
>> [vdsm@node03:/rhev/data-center/mnt/glusterSD/node01.dc-dus.dalason.net:_ssd__storage/fec2eb5e-21b5-496b-9ea5-f718b2cb5556/images/4a55b9c0-d550-4ecb-8dd1-cc1f24f2c7ac]
>>
>>>> $ date ; id ; dd if=5fca6d0e-e320-425b-a89a-f80563461add | pv  | dd
>>>> of=/dev/null
>>>> Mon  3 Feb 11:39:16 CET 2020
>>>> uid=36(vdsm) gid=36(kvm) groups=36(kvm),107(qemu),179(sanlock)
>>>> context=unconfined_u:unconfined_r:unconfined_t:s0-s0:c0.c1023
>>>> dd: error reading ‘5fca6d0e-e320-425b-a89a-f80563461add’: Permission
>>
>>>> denied
>>>> 131072+0 records in
>>>> 131072+0 records out
>>>> 67108864 bytes (67 MB) copied, 0.148846 s, 451 MB/s
>>>>     64MiB 0:00:00 [ 427MiB/s] [  <=>
>>>>        ]
>>>> 131072+0 records in
>>>> 131072+0 records out
>>>> 67108864 bytes (67 MB) copied, 0.149589 s, 449 MB/s
>>>>
>>>>
>>>> try same as root:
>>>>
>>>>
>> [root@node03:/rhev/data-center/mnt/glusterSD/node01.dc-dus.dalason.net:_ssd__storage/fec2eb5e-21b5-496b-9ea5-f718b2cb5556/images/4a55b9c0-d550-4ecb-8dd1-cc1f24f2c7ac]
>>
>>>> # date ; id ; dd if=5fca6d0e-e320-425b-a89a-f80563461add | pv  | dd
>>>> of=/dev/null
>>>> Mon  3 Feb 11:39:33 CET 2020
>>>> uid=0(root) gid=0(root) groups=0(root)
>>>> context=unconfined_u:unconfined_r:unconfined_t:s0-s0:c0.c1023
>>>>     50GiB 0:03:06 [ 274MiB/s] [                              <=>
>>>>        ]
>>>> 104857600+0 records in
>>>> 104857600+0 records out
>>>> 53687091200 bytes (54 GB) copied, 186.501 s, 288 MB/s
>>>> 104857600+0 records in
>>>> 104857600+0 records out
>>>> 53687091200 bytes (54 GB) copied, 186.502 s, 288 MB/s
>>>>
>>>>
>>>> Followed by another vdsm dd test:
>>>>
>>>>
>> [vdsm@node03:/rhev/data-center/mnt/glusterSD/node01.dc-dus.dalason.net:_ssd__storage/fec2eb5e-21b5-496b-9ea5-f718b2cb5556/images/4a55b9c0-d550-4ecb-8dd1-cc1f24f2c7ac]
>>
>>>> $ date ; id ; dd if=5fca6d0e-e320-425b-a89a-f80563461add | pv  | dd
>>>> of=/dev/null
>>>> Mon  3 Feb 11:42:46 CET 2020
>>>> uid=36(vdsm) gid=36(kvm) groups=36(kvm),107(qemu),179(sanlock)
>>>> context=unconfined_u:unconfined_r:unconfined_t:s0-s0:c0.c1023
>>>>     50GiB 0:02:56 [ 290MiB/s] [           <=>      ]
>>>> 104857600+0 records in
>>>> 104857600+0 records out
>>>> 53687091200 bytes (54 GB) copied, 176.189 s, 305 MB/s
>>>> 104857600+0 records in
>>>> 104857600+0 records out
>>>> 53687091200 bytes (54 GB) copied, 176.19 s, 305 MB/s
>>>>
>>>> So it's a permission problem (access denied) unless root loads it
>> first?
>>>> Strange: doing things like file & stat work; I can even cat the meta
>>
>>>> file (small text file). Seems only the disk images (or large files?)
>>
>>>> are affected.
>>>>
>>>> huh!?
>>>>
>>>
>
> What version of gluster  are  you running  ?
> Have you tried  the  solution with :
> 1. Run fake setfacl ?
> Or killing  brick processes  and start the volume with the 'force' option ?
>
> I saw  in your  brick output your selinux context is 'unconfined_u' ... So check for labeling.
> Still,  it looks like my ACL issue.
>
> Best Regards,
> Strahil Nikolov
>

--
with kind regards,
mit freundlichen Gruessen,

Christian Reiss
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-leave@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/
List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/3ZYVQJQR7SFLOG2SUGKP7QIGPDJDGC4D/