As I mentioned in the slack, the safest approach is to:
1. Reduce the volume to replica 1 (there is no need to keep the arbiter until
resynchronization
gluster volume remove-brick VOLUME replica
1 beclovkvma02.bec.net:/data/brick2/brick2 beclovkvma03.bec.net:/data/brick1/brick2 beclovkvma02.bec.net:/data/brick3/brick3
beclovkvma03.bec.net:/data/brick1/brick3 beclovkvma02.bec.net:/data/brick4/brick4
beclovkvma03.bec.net:/data/brick1/brick4 beclovkvma02.bec.net:/data/brick5/brick5
beclovkvma03.bec.net:/data/brick1/brick5 beclovkvma02.bec.net:/data/brick6/brick6
beclovkvma03.bec.net:/data/brick1/brick6 beclovkvma02.bec.net:/data/brick7/brick7
beclovkvma03.bec.net:/data/brick1/brick7 beclovkvma02.bec.net:/data/brick8/brick8
beclovkvma03.bec.net:/data/brick1/brick8 force
Note: I might have missed a brick, so verify that you are selecting all bricks for the
arbiter and beclovkvma02
2. Remove the broken nodegluster peer detach
beclovkvma02.bec.net force
3. Add the freshly installed host:gluster peer probe
beclovkvma04.bec.net
4. Umount all bricks on the arbiter.Then reformat them:mkfs.xfs -f -i size=512
/path/to/each/arbiter/brick/LV
5. Check if fstab is using UUID and if yes -> update with the /dev/VG/LV or with the
new UUIDs (blkid should help)
6. Mount all bricks on the arbiter - no errors should be reported:mount -a
7. Umount , reformat and remount all bricks on
beclovkvma04.bec.net . Don't forget to
check the fstab. 'mount -a' is your first friend
8. Readd the bricks to the volume. Order is important (first 04, then arbiter, 04,
arbiter...)
gluster volume brick-add VOLUME replica 3 arbiter1
beclovkvma04.bec.net:/data/brick2/brick2 beclovkvma03.bec.net:/data/brick1/brick2 beclovkvma04.bec.net:/data/brick3/brick3 beclovkvma03.bec.net:/data/brick1/brick3 beclovkvma04.bec.net:/data/brick4/brick4 beclovkvma03.bec.net:/data/brick1/brick4 beclovkvma04.bec.net:/data/brick5/brick5 beclovkvma03.bec.net:/data/brick1/brick5 beclovkvma04.bec.net:/data/brick6/brick6 beclovkvma03.bec.net:/data/brick1/brick6 beclovkvma04.bec.net:/data/brick7/brick7 beclovkvma03.bec.net:/data/brick1/brick7 beclovkvma04.bec.net:/data/brick8/brick8 beclovkvma03.bec.net:/data/brick1/brick8
9. Trigger the full heal:gluster volume heal VOLUME full
10. If your bricks are high performant and you need to speed up the healing you can
increase these volume settings:
- cluster.shd-max-threads- cluster.shd-wait-qlenght
Best Regards,Strahil Nikolov
On Fri, Nov 12, 2021 at 8:21, dhanaraj.ramesh--- via Users<users(a)ovirt.org> wrote:
wanted to remove
beclovkvma02.bec.net as the node was dead, now I reinstalled this node
and trying to add as 4th node -
beclovkvma04.bec.net however since the system UUID is same
Im not able to add the node in ovirt gluster..
_______________________________________________
Users mailing list -- users(a)ovirt.org
To unsubscribe send an email to users-leave(a)ovirt.org
Privacy Statement:
https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct:
https://www.ovirt.org/community/about/community-guidelines/
List Archives:
https://lists.ovirt.org/archives/list/users@ovirt.org/message/XXY6FD7G6PU...