Appreciate the updates you've been posting. It's concerning
to me as a
Gluster user as well. It would be nice to figure out what happened
here.
On Thu, Feb 6, 2020 at 11:43 AM Christian Reiss
<email(a)christian-reiss.de>
wrote:
> Hey,
>
> For prosperity: Sadly the only way to fix this was to re-init (wipe)
> gluster and start from scratch.
>
> -Chris.
>
> On 03/02/2020 19:23, Strahil Nikolov wrote:
> > On February 3, 2020 2:29:55 PM GMT+02:00, Christian Reiss <
> email(a)christian-reiss.de> wrote:
> >> Ugh,
> >>
> >> disregarding off all previous stamenets:
> >>
> >> new findinds: vdsm user can NOT read files larger than 64mb. Root
can.
> >>
> >> [vdsm@node02:/rhev/data-cente[...]c51d8a18370] $ for i in 60 62 64
66
> >> 68
> >> ; do dd if=/dev/urandom of=file-$i bs=1M count=$i ; done
> >>
> >> [vdsm@node03:/rhev/data-cente[...]c51d8a18370] $ for i in 60 62 64
66
> >> 68
> >> ; do echo $i ; dd if=file-$i of=/dev/null ; done
> >> 60
> >> 122880+0 records in
> >> 122880+0 records out
> >> 62914560 bytes (63 MB) copied, 0.15656 s, 402 MB/s
> >> 62
> >> 126976+0 records in
> >> 126976+0 records out
> >> 65011712 bytes (65 MB) copied, 0.172463 s, 377 MB/s
> >> 64
> >> 131072+0 records in
> >> 131072+0 records out
> >> 67108864 bytes (67 MB) copied, 0.180701 s, 371 MB/s
> >> 66
> >> dd: error reading ‘file-66’: Permission denied
> >> 131072+0 records in
> >> 131072+0 records out
> >> 67108864 bytes (67 MB) copied, 0.105236 s, 638 MB/s
> >> 68
> >> dd: error reading ‘file-68’: Permission denied
> >> 131072+0 records in
> >> 131072+0 records out
> >> 67108864 bytes (67 MB) copied, 0.17046 s, 394 MB/s
> >>
> >>
> >> The files appeared instantly on all nodes, Writing large files
work,
> >> however. Writing large files seem to work.
> >>
> >> I think this is the core issue.
> >>
> >>
> >> On 03/02/2020 12:22, Christian Reiss wrote:
> >>> Further findings:
> >>>
> >>> - modified data gets written to local node, not across gluster.
> >>> - vdsm user can create _new_ files on the cluster, this gets
synced
> >>> immediatly.
> >>> - vdsm can modify, across all nodes newly created files, changes
> >> apply
> >>> immediately.
> >>>
> >>> I think vdsm user can not modify already existing files over the
> >>> gluster. Something selinux?
> >>>
> >>> -Chris.
> >>>
> >>> On 03/02/2020 11:46, Christian Reiss wrote:
> >>>> Hey,
> >>>>
> >>>> I think I am barking up the right tree with something (else)
here;
> >>>> Note the timestamps & id's:
> >>>>
> >>>>
> >>>> dd'ing a disk image as vdsm user, try 1:
> >>>>
> >>>>
> >>
[vdsm@node03:/rhev/data-center/mnt/glusterSD/node01.dc-dus.dalason.net:
>
_ssd__storage/fec2eb5e-21b5-496b-9ea5-f718b2cb5556/images/4a55b9c0-d550-4ecb-8dd1-cc1f24f2c7ac]
> >>
> >>>> $ date ; id ; dd if=5fca6d0e-e320-425b-a89a-f80563461add | pv |
dd
> >>>> of=/dev/null
> >>>> Mon 3 Feb 11:39:13 CET 2020
> >>>> uid=36(vdsm) gid=36(kvm) groups=36(kvm),107(qemu),179(sanlock)
> >>>> context=unconfined_u:unconfined_r:unconfined_t:s0-s0:c0.c1023
> >>>> dd: error reading ‘5fca6d0e-e320-425b-a89a-f80563461add’:
Permission
> >>
> >>>> denied
> >>>> 131072+0 records in
> >>>> 131072+0 records out
> >>>> 67108864 bytes (67 MB) copied, 0.169465 s, 396 MB/s
> >>>> 64MiB 0:00:00 [ 376MiB/s] [ <=>
> >>>> ]
> >>>> 131072+0 records in
> >>>> 131072+0 records out
> >>>> 67108864 bytes (67 MB) copied, 0.171726 s, 391 MB/s
> >>>>
> >>>>
> >>>> try 2, directly afterward:
> >>>>
> >>>>
> >>
[vdsm@node03:/rhev/data-center/mnt/glusterSD/node01.dc-dus.dalason.net:
>
_ssd__storage/fec2eb5e-21b5-496b-9ea5-f718b2cb5556/images/4a55b9c0-d550-4ecb-8dd1-cc1f24f2c7ac]
> >>
> >>>> $ date ; id ; dd if=5fca6d0e-e320-425b-a89a-f80563461add | pv |
dd
> >>>> of=/dev/null
> >>>> Mon 3 Feb 11:39:16 CET 2020
> >>>> uid=36(vdsm) gid=36(kvm) groups=36(kvm),107(qemu),179(sanlock)
> >>>> context=unconfined_u:unconfined_r:unconfined_t:s0-s0:c0.c1023
> >>>> dd: error reading ‘5fca6d0e-e320-425b-a89a-f80563461add’:
Permission
> >>
> >>>> denied
> >>>> 131072+0 records in
> >>>> 131072+0 records out
> >>>> 67108864 bytes (67 MB) copied, 0.148846 s, 451 MB/s
> >>>> 64MiB 0:00:00 [ 427MiB/s] [ <=>
> >>>> ]
> >>>> 131072+0 records in
> >>>> 131072+0 records out
> >>>> 67108864 bytes (67 MB) copied, 0.149589 s, 449 MB/s
> >>>>
> >>>>
> >>>> try same as root:
> >>>>
> >>>>
> >>
[root@node03:/rhev/data-center/mnt/glusterSD/node01.dc-dus.dalason.net:
>
_ssd__storage/fec2eb5e-21b5-496b-9ea5-f718b2cb5556/images/4a55b9c0-d550-4ecb-8dd1-cc1f24f2c7ac]
> >>
> >>>> # date ; id ; dd if=5fca6d0e-e320-425b-a89a-f80563461add | pv |
dd
> >>>> of=/dev/null
> >>>> Mon 3 Feb 11:39:33 CET 2020
> >>>> uid=0(root) gid=0(root) groups=0(root)
> >>>> context=unconfined_u:unconfined_r:unconfined_t:s0-s0:c0.c1023
> >>>> 50GiB 0:03:06 [ 274MiB/s] [
<=>
> >>>> ]
> >>>> 104857600+0 records in
> >>>> 104857600+0 records out
> >>>> 53687091200 bytes (54 GB) copied, 186.501 s, 288 MB/s
> >>>> 104857600+0 records in
> >>>> 104857600+0 records out
> >>>> 53687091200 bytes (54 GB) copied, 186.502 s, 288 MB/s
> >>>>
> >>>>
> >>>> Followed by another vdsm dd test:
> >>>>
> >>>>
> >>
[vdsm@node03:/rhev/data-center/mnt/glusterSD/node01.dc-dus.dalason.net:
>
_ssd__storage/fec2eb5e-21b5-496b-9ea5-f718b2cb5556/images/4a55b9c0-d550-4ecb-8dd1-cc1f24f2c7ac]
> >>
> >>>> $ date ; id ; dd if=5fca6d0e-e320-425b-a89a-f80563461add | pv |
dd
> >>>> of=/dev/null
> >>>> Mon 3 Feb 11:42:46 CET 2020
> >>>> uid=36(vdsm) gid=36(kvm) groups=36(kvm),107(qemu),179(sanlock)
> >>>> context=unconfined_u:unconfined_r:unconfined_t:s0-s0:c0.c1023
> >>>> 50GiB 0:02:56 [ 290MiB/s] [ <=> ]
> >>>> 104857600+0 records in
> >>>> 104857600+0 records out
> >>>> 53687091200 bytes (54 GB) copied, 176.189 s, 305 MB/s
> >>>> 104857600+0 records in
> >>>> 104857600+0 records out
> >>>> 53687091200 bytes (54 GB) copied, 176.19 s, 305 MB/s
> >>>>
> >>>> So it's a permission problem (access denied) unless root loads
it
> >> first?
> >>>> Strange: doing things like file & stat work; I can even cat the
meta
> >>
> >>>> file (small text file). Seems only the disk images (or large
files?)
> >>
> >>>> are affected.
> >>>>
> >>>> huh!?
> >>>>
> >>>
> >
> > What version of gluster are you running ?
> > Have you tried the solution with :
> > 1. Run fake setfacl ?
> > Or killing brick processes and start the volume with the 'force'
> option ?
> >
> > I saw in your brick output your selinux context is 'unconfined_u'
...
> So check for labeling.
> > Still, it looks like my ACL issue.
> >
> > Best Regards,
> > Strahil Nikolov
> >
>
> --
> with kind regards,
> mit freundlichen Gruessen,
>
> Christian Reiss
> _______________________________________________
> Users mailing list -- users(a)ovirt.org
> To unsubscribe send an email to users-leave(a)ovirt.org
> Privacy Statement:
https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct:
>
https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
>
https://lists.ovirt.org/archives/list/users@ovirt.org/message/3ZYVQJQR7SF...
>
Hi Jayme,
It's the ACL bug that hit me some time ago.
It was supposed to be fixed in 6.6 , yet it isn't (even 7.1 & 7.2 is affected).
For some reason , gluster looses all acl data and thus only root can access the gluster
shards.
Best Regards,
Strahil Nikolov