On Wed, Mar 3, 2021, 19:13 <souvaliotimaria@mail.com> wrote:
Hello,

Thank you very much for your reply.

I get the following from the below gluster commands:

        [root@ov-no1 ~]# gluster volume heal engine info split-brain
                Brick ov-no1.ariadne-t.local:/gluster_bricks/engine/engine
                Status: Connected
                Number of entries in split-brain: 0

                Brick ov-no2.ariadne-t.local:/gluster_bricks/engine/engine
                Status: Connected
                Number of entries in split-brain: 0

                Brick ov-no3.ariadne-t.local:/gluster_bricks/engine/engine
                Status: Connected
                Number of entries in split-brain: 0


        [root@ov-no1 ~]# gluster volume heal engine info summary
                Brick ov-no1.ariadne-t.local:/gluster_bricks/engine/engine
                Status: Connected
                Total Number of entries: 1
                Number of entries in heal pending: 1
                Number of entries in split-brain: 0
                Number of entries possibly healing: 0

                Brick ov-no2.ariadne-t.local:/gluster_bricks/engine/engine
                Status: Connected
                Total Number of entries: 1
                Number of entries in heal pending: 1
                Number of entries in split-brain: 0
                Number of entries possibly healing: 0

                Brick ov-no3.ariadne-t.local:/gluster_bricks/engine/engine
                Status: Connected
                Total Number of entries: 1
                Number of entries in heal pending: 1
                Number of entries in split-brain: 0
                Number of entries possibly healing: 0


        [root@ov-no1 ~]# gluster volume info
                Volume Name: data
                        Type: Replicate
                        Volume ID: 6c7bb2e4-ed35-4826-81f6-34fcd2d0a984
                        Status: Started
                        Snapshot Count: 0
                        Number of Bricks: 1 x (2 + 1) = 3
                        Transport-type: tcp
                        Bricks:
                        Brick1: ov-no1.ariadne-t.local:/gluster_bricks/data/data
                        Brick2: ov-no2.ariadne-t.local:/gluster_bricks/data/data
                        Brick3: ov-no3.ariadne-t.local:/gluster_bricks/data/data (arbiter)
                        Options Reconfigured:
                        performance.client-io-threads: on
                        nfs.disable: on
                        transport.address-family: inet
                        performance.strict-o-direct: on
                        performance.quick-read: off
                        performance.read-ahead: off
                        performance.io-cache: off
                        performance.low-prio-threads: 32
                        network.remote-dio: off
                        cluster.eager-lock: enable
                        cluster.quorum-type: auto
                        cluster.server-quorum-type: server
                        cluster.data-self-heal-algorithm: full
                        cluster.locking-scheme: granular
                        cluster.shd-max-threads: 8
                        cluster.shd-wait-qlength: 10000
                        features.shard: on
                        user.cifs: off
                        cluster.choose-local: off
                        client.event-threads: 4
                        server.event-threads: 4
                        network.ping-timeout: 30
                        storage.owner-uid: 36
                        storage.owner-gid: 36
                        cluster.granular-entry-heal: enable

                Volume Name: engine
                        Type: Replicate
                        Volume ID: 7173c827-309f-4e84-a0da-6b2b8eb50264
                        Status: Started
                        Snapshot Count: 0
                        Number of Bricks: 1 x 3 = 3
                        Transport-type: tcp
                        Bricks:
                        Brick1: ov-no1.ariadne-t.local:/gluster_bricks/engine/engine
                        Brick2: ov-no2.ariadne-t.local:/gluster_bricks/engine/engine
                        Brick3: ov-no3.ariadne-t.local:/gluster_bricks/engine/engine
                        Options Reconfigured:
                        performance.client-io-threads: on
                        nfs.disable: on
                        transport.address-family: inet
                        performance.strict-o-direct: on
                        performance.quick-read: off
                        performance.read-ahead: off
                        performance.io-cache: off
                        performance.low-prio-threads: 32
                        network.remote-dio: off
                        cluster.eager-lock: enable
                        cluster.quorum-type: auto
                        cluster.server-quorum-type: server
                        cluster.data-self-heal-algorithm: full
                        cluster.locking-scheme: granular
                        cluster.shd-max-threads: 8
                        cluster.shd-wait-qlength: 10000
                        features.shard: on
                        user.cifs: off
                        cluster.choose-local: off
                        client.event-threads: 4
                        server.event-threads: 4
                        network.ping-timeout: 30
                        storage.owner-uid: 36
                        storage.owner-gid: 36
                        cluster.granular-entry-heal: enable

                Volume Name: vmstore
                        Type: Replicate
                        Volume ID: 29992fc1-3e09-4360-b651-4449fcd32767
                        Status: Started
                        Snapshot Count: 0
                        Number of Bricks: 1 x (2 + 1) = 3
                        Transport-type: tcp
                        Bricks:
                        Brick1: ov-no1.ariadne-t.local:/gluster_bricks/vmstore/vmstore
                        Brick2: ov-no2.ariadne-t.local:/gluster_bricks/vmstore/vmstore
                        Brick3: ov-no3.ariadne-t.local:/gluster_bricks/vmstore/vmstore (arbiter)
                        Options Reconfigured:
                        performance.client-io-threads: on
                        nfs.disable: on
                        transport.address-family: inet
                        performance.strict-o-direct: on
                        performance.quick-read: off
                        performance.read-ahead: off
                        performance.io-cache: off
                        performance.low-prio-threads: 32
                        network.remote-dio: off
                        cluster.eager-lock: enable
                        cluster.quorum-type: auto
                        cluster.server-quorum-type: server
                        cluster.data-self-heal-algorithm: full
                        cluster.locking-scheme: granular
                        cluster.shd-max-threads: 8
                        cluster.shd-wait-qlength: 10000
                        features.shard: on
                        user.cifs: off
                        cluster.choose-local: off
                        client.event-threads: 4
                        server.event-threads: 4
                        network.ping-timeout: 30
                        storage.owner-uid: 36
                        storage.owner-gid: 36
                        cluster.granular-entry-heal: enable


        [root@ov-no1 ~]# gluster volume heal engine info
                Brick ov-no1.ariadne-t.local:/gluster_bricks/engine/engine
                Status: Connected
                Number of entries: 0

                Brick ov-no2.ariadne-t.local:/gluster_bricks/engine/engine
                /80f6e393-9718-4738-a14a-64cf43c3d8c2/images/d5de54b6-9f8e-4fba-819b-ebf6780757d2/a48555f4-be23-4467-8a54-400ae7baf9d7
                Status: Connected
                Number of entries: 1

                Brick ov-no3.ariadne-t.local:/gluster_bricks/engine/engine
                /80f6e393-9718-4738-a14a-64cf43c3d8c2/images/d5de54b6-9f8e-4fba-819b-ebf6780757d2/a48555f4-be23-4467-8a54-400ae7baf9d7
                Status: Connected
                Number of entries: 1


However, checking the contents of the above entry in each host I get the following output in which it's shown that the file in the third host has a different date (the current date):


        [root@ov-no1 ~]# ls /rhev/data-center/mnt/glusterSD/ov-no1.ariadne-t.local\:_engine/80f6e393-9718-4738-a14a-64cf43c3d8c2/images/d5de54b6-9f8e-4fba-819b-ebf6780757d2/
                total 4.6G
                drwxr-xr-x. 2 vdsm kvm  149 Sep 11  2019 .
                drwxr-xr-x. 8 vdsm kvm 8.0K Sep 11  2019 ..
                -rw-rw----. 1 vdsm kvm 100G Dec 30 13:20 a48555f4-be23-4467-8a54-400ae7baf9d7
                -rw-rw----. 1 vdsm kvm 1.0M Feb 24 20:50 a48555f4-be23-4467-8a54-400ae7baf9d7.lease
                -rw-r--r--. 1 vdsm kvm  321 Sep 11  2019 a48555f4-be23-4467-8a54-400ae7baf9d7.meta

        [root@ov-no2 ~]# ls /rhev/data-center/mnt/glusterSD/ov-no1.ariadne-t.local\:_engine/80f6e393-9718-4738-a14a-64cf43c3d8c2/images/d5de54b6-9f8e-4fba-819b-ebf6780757d2/
                total 4.6G
                drwxr-xr-x. 2 vdsm kvm  149 Sep 11  2019 .
                drwxr-xr-x. 8 vdsm kvm 8.0K Sep 11  2019 ..
                -rw-rw----. 1 vdsm kvm 100G Dec 30 13:20 a48555f4-be23-4467-8a54-400ae7baf9d7
                -rw-rw----. 1 vdsm kvm 1.0M Feb 24 20:50 a48555f4-be23-4467-8a54-400ae7baf9d7.lease
                -rw-r--r--. 1 vdsm kvm  321 Sep 11  2019 a48555f4-be23-4467-8a54-400ae7baf9d7.meta

        [root@ov-no3 ~]# ls /rhev/data-center/mnt/glusterSD/ov-no1.ariadne-t.local\:_engine/80f6e393-9718-4738-a14a-64cf43c3d8c2/images/d5de54b6-9f8e-4fba-819b-ebf6780757d2/
                total 4.6G
                drwxr-xr-x. 2 vdsm kvm  149 Sep 11  2019 .
                drwxr-xr-x. 8 vdsm kvm 8.0K Sep 11  2019 ..
                -rw-rw----. 1 vdsm kvm 100G Mar  3 18:13 a48555f4-be23-4467-8a54-400ae7baf9d7
                -rw-rw----. 1 vdsm kvm 1.0M Feb 24 20:50 a48555f4-be23-4467-8a54-400ae7baf9d7.lease
                -rw-r--r--. 1 vdsm kvm  321 Sep 11  2019 a48555f4-be23-4467-8a54-400ae7baf9d7.meta

Also, the stat command on each host gives the following:

        [root@ov-no1 ~]# stat /rhev/data-center/mnt/glusterSD/ov-no1.ariadne-t.local\:_engine/80f6e393-9718-4738-a14a-64cf43c3d8c2/images/d5de54b6-9f8e-4fba-819b-ebf6780757d2/a48555f4-be23-4467-8a54-400ae7baf9d7
                  File: ‘/rhev/data-center/mnt/glusterSD/ov-no1.ariadne-t.local:_engine/80f6e393-9718-4738-a14a-64cf43c3d8c2/images/d5de54b6-9f8e-4fba-819b-ebf6780757d2/a48555f4-be23-4467-8a54-400ae7baf9d7’
                  Size: 107374182400    Blocks: 9569291    IO Block: 131072 regular file
                Device: 29h/41d Inode: 10220711633933694927  Links: 1
                Access: (0660/-rw-rw----)  Uid: (   36/    vdsm)   Gid: (   36/     kvm)
                Context: system_u:object_r:fusefs_t:s0
                Access: 2019-09-11 19:08:58.012200046 +0300
                Modify: 2020-12-30 13:20:39.794315096 +0200
                Change: 2020-12-30 13:20:39.794315096 +0200
                 Birth: -

        [root@ov-no2 ~]# stat  /rhev/data-center/mnt/glusterSD/ov-no1.ariadne-t.local\:_engine/80f6e393-9718-4738-a14a-64cf43c3d8c2/images/d5de54b6-9f8e-4fba-819b-ebf6780757d2/a48555f4-be23-4467-8a54-400ae7baf9d7
                  File: ‘/rhev/data-center/mnt/glusterSD/ov-no1.ariadne-t.local:_engine/80f6e393-9718-4738-a14a-64cf43c3d8c2/images/d5de54b6-9f8e-4fba-819b-ebf6780757d2/a48555f4-be23-4467-8a54-400ae7baf9d7’
                Size: 107374182400    Blocks: 9569291    IO Block: 131072 regular file
                Device: 29h/41d Inode: 10220711633933694927  Links: 1
                Access: (0660/-rw-rw----)  Uid: (   36/    vdsm)   Gid: (   36/     kvm)
                Context: system_u:object_r:fusefs_t:s0
                Access: 2019-09-11 19:08:58.012200046 +0300
                Modify: 2020-12-30 13:20:39.794315096 +0200
                Change: 2020-12-30 13:20:39.794315096 +0200
                 Birth: -

        [root@ov-no3 ~]# stat  /rhev/data-center/mnt/glusterSD/ov-no1.ariadne-t.local\:_engine/80f6e393-9718-4738-a14a-64cf43c3d8c2/images/d5de54b6-9f8e-4fba-819b-ebf6780757d2/a48555f4-be23-4467-8a54-400ae7baf9d7
                  File: ‘/rhev/data-center/mnt/glusterSD/ov-no1.ariadne-t.local:_engine/80f6e393-9718-4738-a14a-64cf43c3d8c2/images/d5de54b6-9f8e-4fba-819b-ebf6780757d2/a48555f4-be23-4467-8a54-400ae7baf9d7’
                Size: 107374182400    Blocks: 9569291    IO Block: 131072 regular file
                Device: 29h/41d Inode: 10220711633933694927  Links: 1
                Access: (0660/-rw-rw----)  Uid: (   36/    vdsm)   Gid: (   36/     kvm)
                Context: system_u:object_r:fusefs_t:s0
                Access: 2020-10-02 03:02:51.104699119 +0300
                Modify: 2021-03-03 18:23:07.122575696 +0200
                Change: 2021-03-03 18:23:07.122575696 +0200
                 Birth: -


Should I use the
gluster volume heal <VOLNAME> split-brain source-brick <HOSTNAME:BRICKNAME>
command to initiate the healing process?
Did you try to heal with latest-mtime?
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-leave@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/
List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/KDENIOVUBH662LT7CQYKCBR5RZSIY32H/