Hi @Dominique,
Once you have added the SSD's to your host, you can follow this link[1] to
do a replace host.
Replace host is essentially
1) Preparing your node based on your current healthy nodes in your cluster
2) Post adding the node in the cluster, gluster syncs with the nodes, and
your cluster will be back to healthy.
[1]
https://github.com/gluster/gluster-ansible/blob/master/playbooks/hc-ansib...
Regards,
Prajith,
On Tue, Jun 22, 2021 at 10:56 AM Ritesh Chikatwar <rchikatw(a)redhat.com>
wrote:
Adding Prajith,
Will replace host work in this case..? If yes please update your thoughts
here
On Tue, Jun 22, 2021 at 9:09 AM Strahil Nikolov via Users <users(a)ovirt.org>
wrote:
> I'm not sure about the GUI (but I think it has the option) , but under
> command line you got several options.
>
> 1. Use gluster's remove-brick replica 2 (with flag force)
> and then 'add-brick replica 3'
> 2. Use the old way 'replace-brick'
>
> If you need guidance, please provide the 'gluster volume info <VOLUME>'
.
>
> Best Regards,
> Strahil Nikolov
>
> On Tue, Jun 22, 2021 at 2:01, Dominique D
> <dominique.deschenes(a)gcgenicom.com> wrote:
> yesterday I had a disk failure on my stack of 3 Ovirt 4.4.1 node
>
> on each server I have 3 Bricks (engine, data, vmstore)
>
> brick data 4X600Gig raid0. /dev/gluster_vg_sdb/gluster_lv_data mount
> /gluster_bricks/data
> brick engine 2X1TB raid1 /dev/gluster_vg_sdc/gluster_lv_engine mount
> /gluster_bricks/engine
> brick vmstore 2X1TB raid1 /dev/gluster_vg_sdc/gluster_lv_vmstore mount
> /gluster_bricks/vmstore
>
> Everything was configured by the gui (hyperconverge and hosted-engine)
>
> It is the raid0 of the 2nd server who broke.
>
> all VMs were automatically moved to the other two servers, I haven't lost
> any data.
>
> the host2 is now in maintenance mode.
>
> I am going to buy 4 new SSD disks to replace the 4 disks of the defective
> raid0.
>
> When I'm going to erase the faulty raid0 and create the new raid with the
> new disks on the raid controler, how do I add in ovirt so that they
> resynchronize with the other bricks data?
>
> Status of volume: data
> Gluster process TCP Port RDMA Port Online
> Pid
>
> ------------------------------------------------------------------------------
> Brick 172.16.70.91:/gluster_bricks/data/dat
> a 49153 0 Y
> 79168
> Brick 172.16.70.92:/gluster_bricks/data/dat
> a N/A N/A N N/A
> Brick 172.16.70.93:/gluster_bricks/data/dat
> a 49152 0 Y 3095
> Self-heal Daemon on localhost N/A N/A Y 2528
> Self-heal Daemon on 172.16.70.91 N/A N/A Y
> 225523
> Self-heal Daemon on 172.16.70.93 N/A N/A Y
> 3121
> _______________________________________________
> Users mailing list -- users(a)ovirt.org
> To unsubscribe send an email to users-leave(a)ovirt.org
> Privacy Statement:
https://www.ovirt.org/privacy-policy.html
> oVirt Code of Conduct:
>
https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
>
https://lists.ovirt.org/archives/list/users@ovirt.org/message/P4XAPOA35NE...
>
> _______________________________________________
> Users mailing list -- users(a)ovirt.org
> To unsubscribe send an email to users-leave(a)ovirt.org
> Privacy Statement:
https://www.ovirt.org/privacy-policy.html
> oVirt Code of Conduct:
>
https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
>
https://lists.ovirt.org/archives/list/users@ovirt.org/message/KHBYPWP23TI...
>