Hi Nir,
thanks for your feedback. My guess is that as gluster is setup to use shard - the first
one (the volume file) is OK ,but then the shard is giving an error.
As gluster1=ovirt1 and gluster2=ovirt2 & ovirt3 is the arbiter , I think it's not
SELINUX issue (all nodes are still in permissive - just to be sure that nothing comes up)
on gluster level.
As I can't powerup any VM (in the data_fastX domains), I took more radical step and
enabled Trace log on the Fuse mount and started a VM.
As per the log , I got :[2019-11-23 17:15:16.961636] W [MSGID: 114031]
[client-rpc-fops_v2.c:2634:client4_0_lookup_cbk] 0-data_fast-client-2: remote operation
failed. Path: /.shard/b0af2b81-22cf-482e-9b2f-c431b6449da
e.79 (00000000-0000-0000-0000-000000000000) [Permission denied]
[2019-11-23 17:15:16.961669] D [MSGID: 0] [client-rpc-fops_v2.c:2642:client4_0_lookup_cbk]
0-stack-trace: stack-address: 0x7fbeb4005218, data_fast-client-2 returned -1 error:
Permission denied
[Permission denied]
[2019-11-23 17:15:16.961705] D [MSGID: 0] [afr-common.c:2464:afr_lookup_done]
0-stack-trace: stack-address: 0x7fbeb4005218, data_fast-replicate-0 returned -1 error:
Permission denied [Permissi
on denied]
[2019-11-23 17:15:16.961735] E [MSGID: 133010]
[shard.c:2327:shard_common_lookup_shards_cbk] 0-data_fast-shard: Lookup on shard 79
failed. Base file gfid = b0af2b81-22cf-482e-9b2f-c431b6449dae
[Permission denied]
[2019-11-23 17:15:16.961764] D [MSGID: 0] [shard.c:842:shard_common_failure_unwind]
0-stack-trace: stack-address: 0x7fbeb4005218, data_fast-shard returned -1 error:
Permission denied [Permissi
on denied]
[2019-11-23 17:15:16.961793] D [MSGID: 0] [utime-autogen-fops.c:100:gf_utime_readv_cbk]
0-stack-trace: stack-address: 0x7fbeb4005218, data_fast-utime returned -1 error:
Permission denied [Perm
ission denied]
[2019-11-23 17:15:16.961821] D [MSGID: 0] [defaults.c:1189:default_readv_cbk]
0-stack-trace: stack-address: 0x7fbeb4005218, data_fast-write-behind returned -1 error:
Permission denied [Permiss
ion denied]
[2019-11-23 17:15:16.961848] D [MSGID: 0] [defaults.c:1189:default_readv_cbk]
0-stack-trace: stack-address: 0x7fbeb4005218, data_fast-open-behind returned -1 error:
Permission denied [Permissi
on denied]
[2019-11-23 17:15:16.961876] D [MSGID: 0] [md-cache.c:2010:mdc_readv_cbk] 0-stack-trace:
stack-address: 0x7fbeb4005218, data_fast-md-cache returned -1 error: Permission denied
[Permission deni
ed]
[2019-11-23 17:15:16.961902] D [MSGID: 0] [defaults.c:1189:default_readv_cbk]
0-stack-trace: stack-address: 0x7fbeb4005218, data_fast-io-threads returned -1 error:
Permission denied [Permissio
n denied]
[2019-11-23 17:15:16.961938] W [fuse-bridge.c:2830:fuse_readv_cbk] 0-glusterfs-fuse: 703:
READ => -1 gfid=b0af2b81-22cf-482e-9b2f-c431b6449dae fd=0x7fbeb4017018 (Permission
denied)
This indicates that the shard 79 for gfid b0af2b81-22cf-482e-9b2f-c431b6449dae has to be
checked.
So I first check the brick logs on ovirt1 from the moment when VM was powered
up:[2019-11-23 17:15:16.962013] I [MSGID: 139001]
[posix-acl.c:263:posix_acl_log_permit_denied] 0-data_fast-access-control: client:
CTX_ID:43c4e254-d99e-45b3-a6b2-dba28b9c2269-GRAPH_ID:0-PID:1842
4-HOST:ovirt2.localdomain-PC_NAME:data_fast-client-0-RECON_NO:-0, gfid:
be318638-e8a0-4c6d-977d-7a937aa84806, req(uid:107,gid:107,perm:1,ngrps:4),
ctx(uid:0,gid:0,in-groups:0,perm:000,updated-
fop:INVALID, acl:-) [Permission denied]
[2019-11-23 17:15:16.962113] E [MSGID: 115050]
[server-rpc-fops_v2.c:158:server4_lookup_cbk] 0-data_fast-server: 541: LOOKUP
/.shard/b0af2b81-22cf-482e-9b2f-c431b6449dae.79 (be318638-e8a0-4c6d
-977d-7a937aa84806/b0af2b81-22cf-482e-9b2f-c431b6449dae.79), client:
CTX_ID:43c4e254-d99e-45b3-a6b2-dba28b9c2269-GRAPH_ID:0-PID:18424-HOST:ovirt2.localdomain-PC_NAME:data_fast-client-0-RECON_N
O:-0, error-xlator: data_fast-access-control [Permission denied]
[2019-11-23 17:15:16.962096] I [MSGID: 139001]
[posix-acl.c:263:posix_acl_log_permit_denied] 0-data_fast-access-control: client:
CTX_ID:43c4e254-d99e-45b3-a6b2-dba28b9c2269-GRAPH_ID:0-PID:1842
4-HOST:ovirt2.localdomain-PC_NAME:data_fast-client-0-RECON_NO:-0, gfid:
be318638-e8a0-4c6d-977d-7a937aa84806, req(uid:107,gid:107,perm:1,ngrps:4),
ctx(uid:0,gid:0,in-groups:0,perm:000,updated-
fop:INVALID, acl:-) [Permission denied]
All of the above doesn't give me any clue what is it going about.And the actual
shard:
[root@ovirt1 .shard]# ls -lahRZ b0af2b81-22cf-482e-9b2f-c431b6449dae.79
-rw-rw----. vdsm kvm system_u:object_r:glusterd_brick_t:s0
b0af2b81-22cf-482e-9b2f-c431b6449dae.79
[root@ovirt1 .shard]# getfattr -d -e hex -m. b0af2b81-22cf-482e-9b2f-c431b6449dae.79
# file: b0af2b81-22cf-482e-9b2f-c431b6449dae.79
security.selinux=0x73797374656d5f753a6f626a6563745f723a676c7573746572645f627269636b5f743a733000
trusted.afr.dirty=0x000000000000000000000000
trusted.gfid=0xac073db9c9934d0bb70bb50dd11bee31
trusted.gfid2path.bedc7b97ff07378d=0x62653331383633382d653861302d346336642d393737642d3761393337616138343830362f62306166326238312d323263662d343832652d396232662d6334333162363434396461652e3739
trusted.glusterfs.mdata=0x010000000000000000000000005dc94ff0000000001eb592b1000000005dc94ff0000000001eb592b1000000005dc94ff0000000001eb592b1
[root@ovirt2 .shard]# ls -lahRZ b0af2b81-22cf-482e-9b2f-c431b6449dae.79
-rw-rw----. vdsm kvm system_u:object_r:glusterd_brick_t:s0
b0af2b81-22cf-482e-9b2f-c431b6449dae.79
[root@ovirt2 .shard]# getfattr -d -e hex -m. b0af2b81-22cf-482e-9b2f-c431b6449dae.79
# file: b0af2b81-22cf-482e-9b2f-c431b6449dae.79
security.selinux=0x73797374656d5f753a6f626a6563745f723a676c7573746572645f627269636b5f743a733000
trusted.afr.dirty=0x000000000000000000000000
trusted.gfid=0xac073db9c9934d0bb70bb50dd11bee31
trusted.gfid2path.bedc7b97ff07378d=0x62653331383633382d653861302d346336642d393737642d3761393337616138343830362f62306166326238312d323263662d343832652d396232662d6334333162363434396461652e3739
trusted.glusterfs.mdata=0x010000000000000000000000005dc94ff0000000001eb592b1000000005dc94ff0000000001eb592b1000000005dc94ff0000000001eb592b1
[root@ovirt3 .shard]# ls -lahRZ b0af2b81-22cf-482e-9b2f-c431b6449dae.79
-rw-rw----. vdsm kvm system_u:object_r:glusterd_brick_t:s0
b0af2b81-22cf-482e-9b2f-c431b6449dae.79
[root@ovirt3 .shard]# getfattr -d -e hex -m. b0af2b81-22cf-482e-9b2f-c431b6449dae.79
# file: b0af2b81-22cf-482e-9b2f-c431b6449dae.79
security.selinux=0x73797374656d5f753a6f626a6563745f723a676c7573746572645f627269636b5f743a733000
trusted.afr.dirty=0x000000000000000000000000
trusted.gfid=0xac073db9c9934d0bb70bb50dd11bee31
trusted.gfid2path.bedc7b97ff07378d=0x62653331383633382d653861302d346336642d393737642d3761393337616138343830362f62306166326238312d323263662d343832652d396232662d6334333162363434396461652e3739
trusted.glusterfs.mdata=0x010000000000000000000000005dc94ff0000000001eb592b1000000005dc94ff0000000001eb592b1000000005dc94ff0000000001eb592b1
Access for user qemu (uid 107/gid 107 -> gluster brick log on ovirt1) is OK:
[root@ovirt1 .shard]# sudo -u qemu dd if=b0af2b81-22cf-482e-9b2f-c431b6449dae.79
of=/dev/null bs=8M iflag=direct status=progress
8+0 прочетени блока
8+0 записани блока
изкопирани са 67108864 байта (67 MB), 0,0130595 s, 5,1 GB/s
[root@ovirt2 .shard]# sudo -u qemu dd if=b0af2b81-22cf-482e-9b2f-c431b6449dae.79
of=/dev/null bs=8M iflag=direct status=progress
8+0 прочетени блока
8+0 записани блока
изкопирани са 67108864 байта (67 MB), 0,0224745 s, 3,0 GB/s
[root@ovirt3 .shard]# sudo -u qemu dd if=b0af2b81-22cf-482e-9b2f-c431b6449dae.79
of=/dev/null bs=8M iflag=direct status=progress
0+0 прочетени блока
0+0 записани блока
изкопирани са 0 байта (0 B), 0,00103977 s, 0,0 kB/s
[root@ovirt1 .shard]# grep 107 /etc/passwd /etc/group
/etc/passwd:qemu:x:107:107:qemu user:/:/sbin/nologin
/etc/group:qemu:x:107:vdsm,sanlock
[root@ovirt2 ~]# grep 107 /etc/passwd /etc/group
/etc/passwd:qemu:x:107:107:qemu user:/:/sbin/nologin
/etc/group:qemu:x:107:vdsm,sanlock
[root@ovirt3 .shard]# grep 107 /etc/passwd /etc/group
/etc/passwd:qemu:x:107:107:qemu user:/:/sbin/nologin
/etc/group:qemu:x:107:vdsm,sanlock
I've created 3 strace files (1 as root and 2 as vdsm) and attached them to this
e-mail:
[root@ovirt2 94f763e9-fd96-4bee-a6b2-31af841a918b]# sudo -u vdsm strace -fvttTyyx -s 4096
-o /tmp/dd-strace-vdsm dd if=5b1d3113-5cca-4582-9029-634b16338a2f of=/dev/null
iflag=direct bs=8M stat
us=progress
dd: грешка при четене на „5b1d3113-5cca-4582-9029-634b16338a2f“: Отказан достъп
8+0 прочетени блока
8+0 записани блока
изкопирани са 67108864 байта (67 MB), 0,118223 s, 568 MB/s
[root@ovirt2 94f763e9-fd96-4bee-a6b2-31af841a918b]# strace -fvttTyyx -s 4096 -o
/tmp/dd-strace-root dd if=5b1d3113-5cca-4582-9029-634b16338a2f of=/dev/null iflag=direct
bs=8M status=progress
изкопирани са 5083496448 байта (5,1 GB), 10,144199 s, 501 MB/s
640+0 прочетени блока
640+0 записани блока
изкопирани са 5368709120 байта (5,4 GB), 10,7444 s, 500 MB/s
[root@ovirt2 94f763e9-fd96-4bee-a6b2-31af841a918b]# sudo -u vdsm strace -fvttTyyx -s 4096
-o /tmp/dd-strace-vdsm-2nd-try dd if=5b1d3113-5cca-4582-9029-634b16338a2f of=/dev/null
iflag=direct bs
=8M status=progress
изкопирани са 4815060992 байта (4,8 GB), 8,056124 s, 598 MB/s
640+0 прочетени блока
640+0 записани блока
изкопирани са 5368709120 байта (5,4 GB), 8,98862 s, 597 MB/s
P.S.: Sorry for the cyrrilic , it's my wife's laptop :). First case gives an error
and reads 67MB, while the second one works as expected.Then vdsm can access the volume
without issue (quite strange)!
В петък, 22 ноември 2019 г., 23:44:27 ч. Гринуич+2, Nir Soffer
<nsoffer(a)redhat.com> написа:
On Fri, Nov 22, 2019 at 10:41 PM Strahil Nikolov <hunter86_bg(a)yahoo.com> wrote:
On Thu, Nov 21, 2019 at 8:20 AM Sahina Bose <sabose(a)redhat.com> wrote:
On Thu, Nov 21, 2019 at 6:03 AM Strahil Nikolov <hunter86_bg(a)yahoo.com> wrote:
Hi All,
another clue in the logs :[2019-11-21 00:29:50.536631] W [MSGID: 114031]
[client-rpc-fops_v2.c:2634:client4_0_lookup_cbk] 0-data_fast-client-1: remote operation
failed. Path: /.shard/b0af2b81-22cf-482e-9b2f-c431b6449dae.79
(00000000-0000-0000-0000-000000000000) [Permission denied][2019-11-21 00:29:50.536798] W
[MSGID: 114031] [client-rpc-fops_v2.c:2634:client4_0_lookup_cbk] 0-data_fast-client-0:
remote operation failed. Path: /.shard/b0af2b81-22cf-482e-9b2f-c431b6449dae.79
(00000000-0000-0000-0000-000000000000) [Permission denied][2019-11-21 00:29:50.536959] W
[MSGID: 114031] [client-rpc-fops_v2.c:2634:client4_0_lookup_cbk] 0-data_fast-client-2:
remote operation failed. Path: /.shard/b0af2b81-22cf-482e-9b2f-c431b6449dae.79
(00000000-0000-0000-0000-000000000000) [Permission denied][2019-11-21 00:29:50.537007] E
[MSGID: 133010] [shard.c:2327:shard_common_lookup_shards_cbk] 0-data_fast-shard: Lookup on
shard 79 failed. Base file gfid = b0af2b81-22cf-482e-9b2f-c431b6449dae [Permission
denied][2019-11-21 00:29:50.537066] W [fuse-bridge.c:2830:fuse_readv_cbk]
0-glusterfs-fuse: 12458: READ => -1 gfid=b0af2b81-22cf-482e-9b2f-c431b6449dae
fd=0x7fc63c00fe18 (Permission denied)[2019-11-21 00:30:01.177665] I [MSGID: 133022]
[shard.c:3674:shard_delete_shards] 0-data_fast-shard: Deleted shards of
gfid=eb103fbf-80dc-425d-882f-1e4efe510db5 from backend[2019-11-21 00:30:13.132756] W
[MSGID: 114031] [client-rpc-fops_v2.c:2634:client4_0_lookup_cbk] 0-data_fast-client-0:
remote operation failed. Path: /.shard/17c663c2-f582-455b-b806-3b9d01fb2c6c.79
(00000000-0000-0000-0000-000000000000) [Permission denied][2019-11-21 00:30:13.132824] W
[MSGID: 114031] [client-rpc-fops_v2.c:2634:client4_0_lookup_cbk] 0-data_fast-client-1:
remote operation failed. Path: /.shard/17c663c2-f582-455b-b806-3b9d01fb2c6c.79
(00000000-0000-0000-0000-000000000000) [Permission denied][2019-11-21 00:30:13.133217] W
[MSGID: 114031] [client-rpc-fops_v2.c:2634:client4_0_lookup_cbk] 0-data_fast-client-2:
remote operation failed. Path: /.shard/17c663c2-f582-455b-b806-3b9d01fb2c6c.79
(00000000-0000-0000-0000-000000000000) [Permission denied][2019-11-21 00:30:13.133238] E
[MSGID: 133010] [shard.c:2327:shard_common_lookup_shards_cbk] 0-data_fast-shard: Lookup on
shard 79 failed. Base file gfid = 17c663c2-f582-455b-b806-3b9d01fb2c6c [Permission
denied][2019-11-21 00:30:13.133264] W [fuse-bridge.c:2830:fuse_readv_cbk]
0-glusterfs-fuse: 12660: READ => -1 gfid=17c663c2-f582-455b-b806-3b9d01fb2c6c
fd=0x7fc63c007038 (Permission denied)[2019-11-21 00:30:38.489449] W [MSGID: 114031]
[client-rpc-fops_v2.c:2634:client4_0_lookup_cbk] 0-data_fast-client-0: remote operation
failed. Path: /.shard/a10a5ae8-108b-4d78-9e65-cca188c27fc4.6
(00000000-0000-0000-0000-000000000000) [Permission denied][2019-11-21 00:30:38.489520] W
[MSGID: 114031] [client-rpc-fops_v2.c:2634:client4_0_lookup_cbk] 0-data_fast-client-1:
remote operation failed. Path: /.shard/a10a5ae8-108b-4d78-9e65-cca188c27fc4.6
(00000000-0000-0000-0000-000000000000) [Permission denied][2019-11-21 00:30:38.489669] W
[MSGID: 114031] [client-rpc-fops_v2.c:2634:client4_0_lookup_cbk] 0-data_fast-client-2:
remote operation failed. Path: /.shard/a10a5ae8-108b-4d78-9e65-cca188c27fc4.6
(00000000-0000-0000-0000-000000000000) [Permission denied][2019-11-21 00:30:38.489717] E
[MSGID: 133010] [shard.c:2327:shard_common_lookup_shards_cbk] 0-data_fast-shard: Lookup on
shard 6 failed. Base file gfid = a10a5ae8-108b-4d78-9e65-cca188c27fc4 [Permission
denied][2019-11-21 00:30:38.489777] W [fuse-bridge.c:2830:fuse_readv_cbk]
0-glusterfs-fuse: 12928: READ => -1 gfid=a10a5ae8-108b-4d78-9e65-cca188c27fc4
fd=0x7fc63c01a058 (Permission denied)
Anyone got an idea why is it happening?I checked user/group and selinux permissions - all
OK
Can you share the commands (and output) used to check this?I first
thought that the file is cached in memory and that's why vdsm user can read the file ,
but the following shows opposite:
[root@ovirt1
94f763e9-fd96-4bee-a6b2-31af841a918b]# lltotal 562145-rw-rw----. 1 vdsm kvm 5368709120 Nov
12 23:29 5b1d3113-5cca-4582-9029-634b16338a2f-rw-rw----. 1 vdsm kvm 1048576 Nov 11
14:11 5b1d3113-5cca-4582-9029-634b16338a2f.lease-rw-r--r--. 1 vdsm kvm 313 Nov 11
14:11 5b1d3113-5cca-4582-9029-634b16338a2f.meta[root@ovirt1
94f763e9-fd96-4bee-a6b2-31af841a918b]#
pwd/rhev/data-center/mnt/glusterSD/gluster1:_data__fast/396604d9-2a9e-49cd-9563-fdc79981f67b/images/94f763e9-fd96-4bee-a6b2-31af841a918b[root@ovirt1
94f763e9-fd96-4bee-a6b2-31af841a918b]# echo 3 > /proc/sys/vm/drop_caches
>I would use iflag=direct instead, no need to mess with caches.
Vdsm always use direct I/O.
[root@ovirt1 94f763e9-fd96-4bee-a6b2-31af841a918b]#
sudo -u vdsm dd if=5b1d3113-5cca-4582-9029-634b16338a2f of=/dev/null bs=4M
status=progressdd: error reading ‘5b1d3113-5cca-4582-9029-634b16338a2f’: Permission
denied
>You got permissions denied...
16+0 records in16+0 records
out67108864 bytes (67 MB) copied, 0.198372 s, 338 MB/s
>And dd continue to read data?!
>I have never seen anything like this.
>It will be helpful to run this with strace:
> strace -t -TT -o dd.strace dd if=vol-id of=/dev/null iflag=direct bs=8M
status=progress
>And share dd.strace.
>Logs in /var/log/glusterfs/exportname.log will contain useful info for this test.
[root@ovirt1 94f763e9-fd96-4bee-a6b2-31af841a918b]# dd
if=5b1d3113-5cca-4582-9029-634b16338a2f of=/dev/null bs=4M status=progress5356126208 bytes
(5.4 GB) copied, 12.061393 s, 444 MB/s1280+0 records in1280+0 records out5368709120 bytes
(5.4 GB) copied, 12.0876 s, 444 MB/s[root@ovirt1 94f763e9-fd96-4bee-a6b2-31af841a918b]#
sudo -u vdsm dd if=5b1d3113-5cca-4582-9029-634b16338a2f of=/dev/null bs=4M
status=progress3598712832 bytes (3.6 GB) copied, 1.000540 s, 3.6 GB/s1280+0 records
in1280+0 records out5368709120 bytes (5.4 GB) copied, 1.47071 s, 3.7 GB/s[root@ovirt1
94f763e9-fd96-4bee-a6b2-31af841a918b]# echo 3 > /proc/sys/vm/drop_caches [root@ovirt1
94f763e9-fd96-4bee-a6b2-31af841a918b]# sudo -u vdsm dd
if=5b1d3113-5cca-4582-9029-634b16338a2f of=/dev/null bs=4M status=progress5171576832 bytes
(5.2 GB) copied, 12.071837 s, 428 MB/s1280+0 records in1280+0 records out5368709120 bytes
(5.4 GB) copied, 12.4873 s, 430 MB/s
As you can see , once root user reads the file -> vdsm user can also do that.
>Smells like issue on gluster side.
I would try this on the hypervisor to check what vdsm/qemu see:
$ ls -lahRZ /rhv/data-center/mnt/glusterSD/gluster-server:_pathI'm
attaching the output of the find I run, but this one should be enough:[root@ovirt1 ~]#
find /rhev/data-center/mnt/glusterSD/*/[0-9]*/images/ -not -user vdsm -print
>A full output of ls -lahRZ, showing user, group, permissions bits,
and selinux label>>of the entire tree will be more useful.
[root@ovirt2
gluster1:_data__fast]# tree
.
├── 396604d9-2a9e-49cd-9563-fdc79981f67b
│ ├── dom_md
│ │ ├── ids
│ │ ├── inbox
│ │ ├── leases
│ │ ├── metadata
│ │ ├── outbox
│ │ └── xleases
│ ├── images
│ │ ├── 225e4c45-0f70-40aa-8d51-7b705f259cd7
│ │ │ ├── 7592ce4c-ecb9-4dbe-ba67-a343ec4cdbc5
│ │ │ ├── 7592ce4c-ecb9-4dbe-ba67-a343ec4cdbc5.lease
│ │ │ ├── 7592ce4c-ecb9-4dbe-ba67-a343ec4cdbc5.meta
│ │ │ ├── b965dd4e-c814-4bb6-8bc5-d1bbb8dc211d
│ │ │ ├── b965dd4e-c814-4bb6-8bc5-d1bbb8dc211d.lease
│ │ │ └── b965dd4e-c814-4bb6-8bc5-d1bbb8dc211d.meta
│ │ ├── 25107759-f56b-4d94-872e-0f9b145886d1
│ │ │ ├── 571a5eb4-eb92-479c-8305-79eaa0673bb2
│ │ │ ├── 571a5eb4-eb92-479c-8305-79eaa0673bb2.lease
│ │ │ ├── 571a5eb4-eb92-479c-8305-79eaa0673bb2.meta
│ │ │ └── 571a5eb4-eb92-479c-8305-79eaa0673bb2.meta_old
│ │ ├── 2c79e7ab-47cc-4f41-a190-40cd94053d83
│ │ │ ├── 6a6d6ff8-92b2-4a14-b54c-77239e28254d
│ │ │ ├── 6a6d6ff8-92b2-4a14-b54c-77239e28254d.lease
│ │ │ ├── 6a6d6ff8-92b2-4a14-b54c-77239e28254d.meta
│ │ │ └── 6a6d6ff8-92b2-4a14-b54c-77239e28254d.meta_old
│ │ ├── 36fed07b-a08f-4fb9-a273-9dbf0ad1f27b
│ │ │ ├── 44ac76fe-dd0d-471d-8ffe-cb9ad28b07a9
│ │ │ ├── 44ac76fe-dd0d-471d-8ffe-cb9ad28b07a9.lease
│ │ │ └── 44ac76fe-dd0d-471d-8ffe-cb9ad28b07a9.meta
│ │ ├── 7d11479e-1a02-4a74-a9be-14b4e56faaa1
│ │ │ ├── 3e9a41f6-652e-4439-bd24-3e7621c27f4a
│ │ │ ├── 3e9a41f6-652e-4439-bd24-3e7621c27f4a.lease
│ │ │ ├── 3e9a41f6-652e-4439-bd24-3e7621c27f4a.meta
│ │ │ ├── a2b94064-9e34-47c9-8eab-867673c5e48f
│ │ │ ├── a2b94064-9e34-47c9-8eab-867673c5e48f.lease
│ │ │ └── a2b94064-9e34-47c9-8eab-867673c5e48f.meta
│ │ ├── 85f18244-6d5b-4d02-ab66-140bfd2d734b
│ │ │ ├── 0cc4fb07-8a08-4c16-bf3f-93d4462c7108
│ │ │ ├── 0cc4fb07-8a08-4c16-bf3f-93d4462c7108.lease
│ │ │ ├── 0cc4fb07-8a08-4c16-bf3f-93d4462c7108.meta
│ │ │ ├── 116191b2-99a1-437b-8b7d-ab1927b5f4aa
│ │ │ ├── 116191b2-99a1-437b-8b7d-ab1927b5f4aa.lease
│ │ │ └── 116191b2-99a1-437b-8b7d-ab1927b5f4aa.meta
│ │ ├── 94f763e9-fd96-4bee-a6b2-31af841a918b
│ │ │ ├── 5b1d3113-5cca-4582-9029-634b16338a2f
│ │ │ ├── 5b1d3113-5cca-4582-9029-634b16338a2f.lease
│ │ │ └── 5b1d3113-5cca-4582-9029-634b16338a2f.meta
│ │ ├── 99c3c20c-a7bb-4851-8ae9-0b39c0c86b66
│ │ │ ├── 6ad3bd5e-6b21-4f10-83b4-07126ca14661
│ │ │ ├── 6ad3bd5e-6b21-4f10-83b4-07126ca14661.lease
│ │ │ ├── 6ad3bd5e-6b21-4f10-83b4-07126ca14661.meta
│ │ │ ├── 95c260da-8dfb-46be-9c4a-6a55eb3967d5
│ │ │ ├── 95c260da-8dfb-46be-9c4a-6a55eb3967d5.lease
│ │ │ └── 95c260da-8dfb-46be-9c4a-6a55eb3967d5.meta
│ │ ├── e56eb458-e459-4828-a268-82a9b91c60f8
│ │ │ ├── 5a7ca769-4657-4f65-92bd-72fa13cc2871
│ │ │ ├── 5a7ca769-4657-4f65-92bd-72fa13cc2871.lease
│ │ │ └── 5a7ca769-4657-4f65-92bd-72fa13cc2871.meta
│ │ └── f5336a0f-de34-482e-b437-47f9d276ed85
│ │ ├── 4f218ba9-a5a2-4f2f-ac6d-e5c9217d4ec4
│ │ ├── 4f218ba9-a5a2-4f2f-ac6d-e5c9217d4ec4.lease
│ │ └── 4f218ba9-a5a2-4f2f-ac6d-e5c9217d4ec4.meta
│ └── master
└── __DIRECT_IO_TEST__
14 directories, 51 files
Full find of "data_fastX" domains are attached.
Also, to make sure we don't have selinux issue on the hypervisor,
you can change>selinux to permissive mode:
> setenforce 0
This is the first thing I did and the systems were still in permissive when I tried
again.I'm 99.99% sure it's not selinux.
And then try again. If this was selinux issue the permission denied
issue will disappear.>If this is the case please provide the output of:
>
ausearh -m AVC -ts today
If the issue still exists, we eliminated selinux, and you can enable
it again:
> setenforce 1
[root@ovirt3 ~]# ausearch -m AVC -ts today<no matches>[root@ovirt2 ~]# ausearch -m
AVC -ts today<no matches>[root@ovirt1 ~]# ausearch -m AVC -ts today<no
matches>
So this is not selinux on the hypervisor. I wonder if it can be
selinux on the gluster side?
I have a vague feeling that the issue is related to
gluster v6.5 to 6.6 upgrade which I several days before... So if any logs are needed (or
debug enabled), just mention.
If this is the last change, and it worked before, most likely.
Gluster was patched several days before RC3 , but all the storage-related issues
lead me to that. The strage thing is that the HostedEngine is not affected, nor the
"data" storage domain .
Any ideas appreciated. I would like to find out what has caused this instead of just
taking the shortcut to redeploy the whole setup.
Best Regards,Strahil Nikolov