On Tue, Nov 27, 2018 at 4:05 PM Abhishek Sahni <abhishek@iiserb.ac.in> wrote:
That is amazing, resetting bricks resolved the issue. 

Thanks much Sahina and Kaustav. 

However, Do we have manual steps to recover those bricks.
https://gluster.readthedocs.io/en/latest/release-notes/3.9.0/ 

On Tue, Nov 27, 2018 at 3:57 PM Abhishek Sahni <abhishek@iiserb.ac.in> wrote:
I just enabled it on default cluster and now the volumes are visible. It seems like gluster service was disabled by default on cluster.

On Tue, Nov 27, 2018 at 3:51 PM Sahina Bose <sabose@redhat.com> wrote:


On Tue, Nov 27, 2018 at 3:45 PM Kaustav Majumder <kmajumde@redhat.com> wrote:
I am not sure why ovirt is not showing any volume.
Sahina, is this a bug?

Check if gluster service is enabled on the cluster.
The volumes are managed only if this is true


On Tue, Nov 27, 2018 at 3:10 PM Abhishek Sahni <abhishek.sahni1991@gmail.com> wrote:
Hello Kaustav,

That's weird, I never saw any volumes under the storage tab since installation. I am using HC setup deployed using cockpit console.


Did I miss something?


On Tue, Nov 27, 2018 at 2:50 PM Kaustav Majumder <kmajumde@redhat.com> wrote:
Click on volume for which you want to reset the brick-> under bricks tab select the brick you wan to reset -> once you do you will see the 'Reset Brick' option is active.
Attached is a screenshot -> https://i.imgur.com/QUMSrzt.png

On Tue, Nov 27, 2018 at 2:43 PM Abhishek Sahni <abhishek.sahni1991@gmail.com> wrote:
Thanks Sahina for your response, I am not able to find it on UI, please help me to navigate? and yes I am using oVirt 4.2.6.4-1.

On Tue, Nov 27, 2018 at 12:55 PM Sahina Bose <sabose@redhat.com> wrote:


On Tue, Nov 20, 2018 at 5:56 PM Abhishek Sahni <abhishek.sahni1991@gmail.com> wrote:
Hello Team,

We are running a setup of 3-way replica HC gluster setup configured during the initial deployment from the cockpit console using ansible.

NODE1
  - /dev/sda   (OS)
  - /dev/sdb   ( Gluster Bricks )
       * /gluster_bricks/engine/engine/ 
       * /gluster_bricks/data/data/ 
       * /gluster_bricks/vmstore/vmstore/ 

NODE2 and NODE3 with a similar setup.

There is a mishap that /dev/sdb on NODE2 totally got crashed and now there is nothing inside. However, I have created similar directories after mounting it back i.e.,

       * /gluster_bricks/engine/engine/ 
       * /gluster_bricks/data/data/ 
       * /gluster_bricks/vmstore/vmstore/ 
but it is not yet recovered.

=====================================================
[root@node2 ~]# gluster volume status
Status of volume: data
Gluster process                             TCP Port  RDMA Port  Online  Pid
------------------------------------------------------------------------------
Brick *.*.*.1:/gluster_bricks/data/data  49152     0          Y       11111
Brick *.*.*.2:/gluster_bricks/data/data  N/A       N/A        N       N/A  
Brick *.*.*.3:/gluster_bricks/data/data  49152     0          Y       4303 
Self-heal Daemon on localhost               N/A       N/A        Y       23976
Self-heal Daemon on *.*.*.1              N/A       N/A        Y       27838
Self-heal Daemon on *.*.*.3              N/A       N/A        Y       27424
 
Task Status of Volume data
------------------------------------------------------------------------------
There are no active volume tasks
 
Status of volume: engine
Gluster process                             TCP Port  RDMA Port  Online  Pid
------------------------------------------------------------------------------
Brick *.*.*.1:/gluster_bricks/engine/eng
ine                                         49153     0          Y       11117
Brick *.*.*.2:/gluster_bricks/engine/eng
ine                                         N/A       N/A        N       N/A  
Brick *.*.*.3:/gluster_bricks/engine/eng
ine                                         49153     0          Y       4314 
Self-heal Daemon on localhost               N/A       N/A        Y       23976
Self-heal Daemon on *.*.*.3              N/A       N/A        Y       27424
Self-heal Daemon on *.*.*.1              N/A       N/A        Y       27838
 
Task Status of Volume engine
------------------------------------------------------------------------------
There are no active volume tasks
 
Status of volume: vmstore
Gluster process                             TCP Port  RDMA Port  Online  Pid
------------------------------------------------------------------------------
Brick *.*.*.1:/gluster_bricks/vmstore/vm
store                                       49154     0          Y       21603
Brick *.*.*.2:/gluster_bricks/vmstore/vm
store                                       N/A       N/A        N       N/A  
Brick *.*.*.3:/gluster_bricks/vmstore/vm
store                                       49154     0          Y       26845
Self-heal Daemon on localhost               N/A       N/A        Y       23976
Self-heal Daemon on *.*.*.3              N/A       N/A        Y       27424
Self-heal Daemon on *.*.*.1              N/A       N/A        Y       27838
 
Task Status of Volume vmstore
------------------------------------------------------------------------------
There are no active volume tasks
=============================================================


Can someone please suggest the steps to recover the setup?

I have tried the below workaround but it doesn't help.


 You can reset the brick - if you're on oVirt 4.2.x, there's a UI option in the bricks subtab to do this.



--

ABHISHEK SAHNI

Mob : +91-990-701-5143


_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-leave@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/
List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/WFYUBA4DPHOSAZHGCRV2AAT4JWL4LWWV/


--

ABHISHEK SAHNI

Mob : +91-990-701-5143


_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-leave@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/
List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/GEZZIGBWA6HQ55JV2PSP7VSSYPZQ2JAC/


Thanks,
Kaustav


--

ABHISHEK SAHNI

Mob : +91-990-701-5143




--

KAUSTAV MAJUMDER

ASSOCIATE SOFTWARE ENGINEER

Red Hat India PVT LTD.

kmajumder@redhat.com    M: 08981884037     IM: IRC: kmajumder



--
Thanks,

Abhishek Sahni
Computer Centre
IISER Bhopal


--
Thanks,

Abhishek Sahni
Computer Centre
IISER Bhopal

Thanks,
Kaustav