Thanks Sahina for your response, I am not able to find it on UI, please
help me to navigate? and yes I am using oVirt 4.2.6.4-1.
On Tue, Nov 27, 2018 at 12:55 PM Sahina Bose <sabose(a)redhat.com> wrote:
On Tue, Nov 20, 2018 at 5:56 PM Abhishek Sahni <
abhishek.sahni1991(a)gmail.com> wrote:
> Hello Team,
>
> We are running a setup of 3-way replica HC gluster setup configured
> during the initial deployment from the cockpit console using ansible.
>
> NODE1
> - /dev/sda (OS)
> - /dev/sdb ( Gluster Bricks )
> * /gluster_bricks/engine/engine/
> * /gluster_bricks/data/data/
> * /gluster_bricks/vmstore/vmstore/
>
> NODE2 and NODE3 with a similar setup.
>
> There is a mishap that /dev/sdb on NODE2 totally got crashed and now
> there is nothing inside. However, I have created similar directories after
> mounting it back i.e.,
>
> * /gluster_bricks/engine/engine/
> * /gluster_bricks/data/data/
> * /gluster_bricks/vmstore/vmstore/
> but it is not yet recovered.
>
> =====================================================
> [root@node2 ~]# gluster volume status
> Status of volume: data
> Gluster process TCP Port RDMA Port Online
> Pid
>
> ------------------------------------------------------------------------------
> Brick *.*.*.1:/gluster_bricks/data/data 49152 0 Y
> 11111
> Brick *.*.*.2:/gluster_bricks/data/data N/A N/A N
> N/A
> Brick *.*.*.3:/gluster_bricks/data/data 49152 0 Y
> 4303
> Self-heal Daemon on localhost N/A N/A Y
> 23976
> Self-heal Daemon on *.*.*.1 N/A N/A Y
> 27838
> Self-heal Daemon on *.*.*.3 N/A N/A Y
> 27424
>
> Task Status of Volume data
>
> ------------------------------------------------------------------------------
> There are no active volume tasks
>
> Status of volume: engine
> Gluster process TCP Port RDMA Port Online
> Pid
>
> ------------------------------------------------------------------------------
> Brick *.*.*.1:/gluster_bricks/engine/eng
> ine 49153 0 Y
> 11117
> Brick *.*.*.2:/gluster_bricks/engine/eng
> ine N/A N/A N
> N/A
> Brick *.*.*.3:/gluster_bricks/engine/eng
> ine 49153 0 Y
> 4314
> Self-heal Daemon on localhost N/A N/A Y
> 23976
> Self-heal Daemon on *.*.*.3 N/A N/A Y
> 27424
> Self-heal Daemon on *.*.*.1 N/A N/A Y
> 27838
>
> Task Status of Volume engine
>
> ------------------------------------------------------------------------------
> There are no active volume tasks
>
> Status of volume: vmstore
> Gluster process TCP Port RDMA Port Online
> Pid
>
> ------------------------------------------------------------------------------
> Brick *.*.*.1:/gluster_bricks/vmstore/vm
> store 49154 0 Y
> 21603
> Brick *.*.*.2:/gluster_bricks/vmstore/vm
> store N/A N/A N
> N/A
> Brick *.*.*.3:/gluster_bricks/vmstore/vm
> store 49154 0 Y
> 26845
> Self-heal Daemon on localhost N/A N/A Y
> 23976
> Self-heal Daemon on *.*.*.3 N/A N/A Y
> 27424
> Self-heal Daemon on *.*.*.1 N/A N/A Y
> 27838
>
> Task Status of Volume vmstore
>
> ------------------------------------------------------------------------------
> There are no active volume tasks
> =============================================================
>
>
> Can someone please suggest the steps to recover the setup?
>
> I have tried the below workaround but it doesn't help.
>
>
>
https://lists.gluster.org/pipermail/gluster-users/2013-November/015079.html
>
You can reset the brick - if you're on oVirt 4.2.x, there's a UI option
in the bricks subtab to do this.
>
> --
>
> ABHISHEK SAHNI
> Mob : +91-990-701-5143
>
>
> _______________________________________________
> Users mailing list -- users(a)ovirt.org
> To unsubscribe send an email to users-leave(a)ovirt.org
> Privacy Statement:
https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct:
>
https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
>
https://lists.ovirt.org/archives/list/users@ovirt.org/message/WFYUBA4DPHO...
>
--
ABHISHEK SAHNI
Mob : +91-990-701-5143