I am not sure why ovirt is not showing any volume.
Sahina, is this a bug?
On Tue, Nov 27, 2018 at 3:10 PM Abhishek Sahni <abhishek.sahni1991(a)gmail.com>
wrote:
Hello Kaustav,
That's weird, I never saw any volumes under the storage tab since
installation. I am using HC setup deployed using cockpit console.
https://imgur.com/a/nH9rzK8
Did I miss something?
On Tue, Nov 27, 2018 at 2:50 PM Kaustav Majumder <kmajumde(a)redhat.com>
wrote:
> Click on volume for which you want to reset the brick-> under bricks tab
> select the brick you wan to reset -> once you do you will see the 'Reset
> Brick' option is active.
> Attached is a screenshot ->
https://i.imgur.com/QUMSrzt.png
>
> On Tue, Nov 27, 2018 at 2:43 PM Abhishek Sahni <
> abhishek.sahni1991(a)gmail.com> wrote:
>
>> Thanks Sahina for your response, I am not able to find it on UI, please
>> help me to navigate? and yes I am using oVirt 4.2.6.4-1.
>>
>> On Tue, Nov 27, 2018 at 12:55 PM Sahina Bose <sabose(a)redhat.com> wrote:
>>
>>>
>>>
>>> On Tue, Nov 20, 2018 at 5:56 PM Abhishek Sahni <
>>> abhishek.sahni1991(a)gmail.com> wrote:
>>>
>>>> Hello Team,
>>>>
>>>> We are running a setup of 3-way replica HC gluster setup configured
>>>> during the initial deployment from the cockpit console using ansible.
>>>>
>>>> NODE1
>>>> - /dev/sda (OS)
>>>> - /dev/sdb ( Gluster Bricks )
>>>> * /gluster_bricks/engine/engine/
>>>> * /gluster_bricks/data/data/
>>>> * /gluster_bricks/vmstore/vmstore/
>>>>
>>>> NODE2 and NODE3 with a similar setup.
>>>>
>>>> There is a mishap that /dev/sdb on NODE2 totally got crashed and now
>>>> there is nothing inside. However, I have created similar directories
after
>>>> mounting it back i.e.,
>>>>
>>>> * /gluster_bricks/engine/engine/
>>>> * /gluster_bricks/data/data/
>>>> * /gluster_bricks/vmstore/vmstore/
>>>> but it is not yet recovered.
>>>>
>>>> =====================================================
>>>> [root@node2 ~]# gluster volume status
>>>> Status of volume: data
>>>> Gluster process TCP Port RDMA Port
>>>> Online Pid
>>>>
>>>>
------------------------------------------------------------------------------
>>>> Brick *.*.*.1:/gluster_bricks/data/data 49152 0 Y
>>>> 11111
>>>> Brick *.*.*.2:/gluster_bricks/data/data N/A N/A N
>>>> N/A
>>>> Brick *.*.*.3:/gluster_bricks/data/data 49152 0 Y
>>>> 4303
>>>> Self-heal Daemon on localhost N/A N/A Y
>>>> 23976
>>>> Self-heal Daemon on *.*.*.1 N/A N/A Y
>>>> 27838
>>>> Self-heal Daemon on *.*.*.3 N/A N/A Y
>>>> 27424
>>>>
>>>> Task Status of Volume data
>>>>
>>>>
------------------------------------------------------------------------------
>>>> There are no active volume tasks
>>>>
>>>> Status of volume: engine
>>>> Gluster process TCP Port RDMA Port
>>>> Online Pid
>>>>
>>>>
------------------------------------------------------------------------------
>>>> Brick *.*.*.1:/gluster_bricks/engine/eng
>>>> ine 49153 0 Y
>>>> 11117
>>>> Brick *.*.*.2:/gluster_bricks/engine/eng
>>>> ine N/A N/A N
>>>> N/A
>>>> Brick *.*.*.3:/gluster_bricks/engine/eng
>>>> ine 49153 0 Y
>>>> 4314
>>>> Self-heal Daemon on localhost N/A N/A Y
>>>> 23976
>>>> Self-heal Daemon on *.*.*.3 N/A N/A Y
>>>> 27424
>>>> Self-heal Daemon on *.*.*.1 N/A N/A Y
>>>> 27838
>>>>
>>>> Task Status of Volume engine
>>>>
>>>>
------------------------------------------------------------------------------
>>>> There are no active volume tasks
>>>>
>>>> Status of volume: vmstore
>>>> Gluster process TCP Port RDMA Port
>>>> Online Pid
>>>>
>>>>
------------------------------------------------------------------------------
>>>> Brick *.*.*.1:/gluster_bricks/vmstore/vm
>>>> store 49154 0 Y
>>>> 21603
>>>> Brick *.*.*.2:/gluster_bricks/vmstore/vm
>>>> store N/A N/A N
>>>> N/A
>>>> Brick *.*.*.3:/gluster_bricks/vmstore/vm
>>>> store 49154 0 Y
>>>> 26845
>>>> Self-heal Daemon on localhost N/A N/A Y
>>>> 23976
>>>> Self-heal Daemon on *.*.*.3 N/A N/A Y
>>>> 27424
>>>> Self-heal Daemon on *.*.*.1 N/A N/A Y
>>>> 27838
>>>>
>>>> Task Status of Volume vmstore
>>>>
>>>>
------------------------------------------------------------------------------
>>>> There are no active volume tasks
>>>> =============================================================
>>>>
>>>>
>>>> Can someone please suggest the steps to recover the setup?
>>>>
>>>> I have tried the below workaround but it doesn't help.
>>>>
>>>>
>>>>
https://lists.gluster.org/pipermail/gluster-users/2013-November/015079.html
>>>>
>>>
>>> You can reset the brick - if you're on oVirt 4.2.x, there's a UI
>>> option in the bricks subtab to do this.
>>>
>>>
>>>>
>>>> --
>>>>
>>>> ABHISHEK SAHNI
>>>> Mob : +91-990-701-5143
>>>>
>>>>
>>>> _______________________________________________
>>>> Users mailing list -- users(a)ovirt.org
>>>> To unsubscribe send an email to users-leave(a)ovirt.org
>>>> Privacy Statement:
https://www.ovirt.org/site/privacy-policy/
>>>> oVirt Code of Conduct:
>>>>
https://www.ovirt.org/community/about/community-guidelines/
>>>> List Archives:
>>>>
https://lists.ovirt.org/archives/list/users@ovirt.org/message/WFYUBA4DPHO...
>>>>
>>>
>>
>> --
>>
>> ABHISHEK SAHNI
>> Mob : +91-990-701-5143
>>
>>
>> _______________________________________________
>> Users mailing list -- users(a)ovirt.org
>> To unsubscribe send an email to users-leave(a)ovirt.org
>> Privacy Statement:
https://www.ovirt.org/site/privacy-policy/
>> oVirt Code of Conduct:
>>
https://www.ovirt.org/community/about/community-guidelines/
>> List Archives:
>>
https://lists.ovirt.org/archives/list/users@ovirt.org/message/GEZZIGBWA6H...
>>
>
>
> Thanks,
> Kaustav
>
--
ABHISHEK SAHNI
Mob : +91-990-701-5143
--
KAUSTAV MAJUMDER
ASSOCIATE SOFTWARE ENGINEER
Red Hat India PVT LTD. <
TRIED. TESTED. TRUSTED. <