yes, that is the same option i was asking about. Apologies that i had
mentioned a different name.
So, ovirt will automatically detect it if you select the option 'use
managed gluster volume'. While adding a storage domain after specifying the
host , you could just select the checkbox and that will list all the
volumes managed from ovirt UI + that will fill the mount options for you.
On Fri, Sep 1, 2017 at 6:40 PM, Charles Kozler <ckozleriii(a)gmail.com> wrote:
Are you referring to "Mount Options" - >
http://i.imgur.com/bYfbyzz.png
Then no, but that would explain why it wasnt working :-). I guess I had a
silly assumption that oVirt would have detected it and automatically taken
up the redundancy that was configured inside the replica set / brick
detection.
I will test and let you know
Thanks!
On Fri, Sep 1, 2017 at 8:52 AM, Kasturi Narra <knarra(a)redhat.com> wrote:
> Hi Charles,
>
> One question, while configuring a storage domain you are
> saying "host to use: " node1, then in the connection details you say
> node1:/data. What about the backup-volfile-servers option in the UI while
> configuring storage domain? Are you specifying that too?
>
> Thanks
> kasturi
>
>
> On Fri, Sep 1, 2017 at 5:52 PM, Charles Kozler <ckozleriii(a)gmail.com>
> wrote:
>
>> @ Jim - you have only two data volumes and lost quorum. Arbitrator only
>> stores metadata, no actual files. So yes, you were running in degraded mode
>> so some operations were hindered.
>>
>> @ Sahina - Yes, this actually worked fine for me once I did that.
>> However, the issue I am still facing, is when I go to create a new gluster
>> storage domain (replica 3, hyperconverged) and I tell it "Host to use"
and
>> I select that host. If I fail that host, all VMs halt. I do not recall this
>> in 3.6 or early 4.0. This to me makes it seem like this is "pinning" a
node
>> to a volume and vice versa like you could, for instance, for a singular
>> hyperconverged to ex: export a local disk via NFS and then mount it via
>> ovirt domain. But of course, this has its caveats. To that end, I am using
>> gluster replica 3, when configuring it I say "host to use: " node 1,
then
>> in the connection details I give it node1:/data. I fail node1, all VMs
>> halt. Did I miss something?
>>
>> On Fri, Sep 1, 2017 at 2:13 AM, Sahina Bose <sabose(a)redhat.com> wrote:
>>
>>> To the OP question, when you set up a gluster storage domain, you need
>>> to specify backup-volfile-servers=<server2>:<server3> where
server2
>>> and server3 also have bricks running. When server1 is down, and the volume
>>> is mounted again - server2 or server3 are queried to get the gluster
>>> volfiles.
>>>
>>> @Jim, if this does not work, are you using 4.1.5 build with libgfapi
>>> access? If not, please provide the vdsm and gluster mount logs to analyse
>>>
>>> If VMs go to paused state - this could mean the storage is not
>>> available. You can check "gluster volume status <volname>" to
see if
>>> atleast 2 bricks are running.
>>>
>>> On Fri, Sep 1, 2017 at 11:31 AM, Johan Bernhardsson <johan(a)kafit.se>
>>> wrote:
>>>
>>>> If gluster drops in quorum so that it has less votes than it should it
>>>> will stop file operations until quorum is back to normal.If i rember it
>>>> right you need two bricks to write for quorum to be met and that the
>>>> arbiter only is a vote to avoid split brain.
>>>>
>>>>
>>>> Basically what you have is a raid5 solution without a spare. And when
>>>> one disk dies it will run in degraded mode. And some raid systems will
stop
>>>> the raid until you have removed the disk or forced it to run anyway.
>>>>
>>>> You can read up on it here:
https://gluster.readthed
>>>> ocs.io/en/latest/Administrator%20Guide/arbiter-volumes-and-quorum/
>>>>
>>>> /Johan
>>>>
>>>> On Thu, 2017-08-31 at 22:33 -0700, Jim Kusznir wrote:
>>>>
>>>> Hi all:
>>>>
>>>> Sorry to hijack the thread, but I was about to start essentially the
>>>> same thread.
>>>>
>>>> I have a 3 node cluster, all three are hosts and gluster nodes
>>>> (replica 2 + arbitrar). I DO have the
mnt_options=backup-volfile-servers=
>>>> set:
>>>>
>>>> storage=192.168.8.11:/engine
>>>> mnt_options=backup-volfile-servers=192.168.8.12:192.168.8.13
>>>>
>>>> I had an issue today where 192.168.8.11 went down. ALL VMs
>>>> immediately paused, including the engine (all VMs were running on
>>>> host2:192.168.8.12). I couldn't get any gluster stuff working until
host1
>>>> (192.168.8.11) was restored.
>>>>
>>>> What's wrong / what did I miss?
>>>>
>>>> (this was set up "manually" through the article on setting up
>>>> self-hosted gluster cluster back when 4.0 was new..I've upgraded it
to 4.1
>>>> since).
>>>>
>>>> Thanks!
>>>> --Jim
>>>>
>>>>
>>>> On Thu, Aug 31, 2017 at 12:31 PM, Charles Kozler
<ckozleriii(a)gmail.com
>>>> > wrote:
>>>>
>>>> Typo..."Set it up and then failed that **HOST**"
>>>>
>>>> And upon that host going down, the storage domain went down. I only
>>>> have hosted storage domain and this new one - is this why the DC went
down
>>>> and no SPM could be elected?
>>>>
>>>> I dont recall this working this way in early 4.0 or 3.6
>>>>
>>>> On Thu, Aug 31, 2017 at 3:30 PM, Charles Kozler
<ckozleriii(a)gmail.com>
>>>> wrote:
>>>>
>>>> So I've tested this today and I failed a node. Specifically, I setup
a
>>>> glusterfs domain and selected "host to use: node1". Set it up
and then
>>>> failed that VM
>>>>
>>>> However, this did not work and the datacenter went down. My engine
>>>> stayed up, however, it seems configuring a domain to pin to a host to
use
>>>> will obviously cause it to fail
>>>>
>>>> This seems counter-intuitive to the point of glusterfs or any
>>>> redundant storage. If a single host has to be tied to its function, this
>>>> introduces a single point of failure
>>>>
>>>> Am I missing something obvious?
>>>>
>>>> On Thu, Aug 31, 2017 at 9:43 AM, Kasturi Narra <knarra(a)redhat.com>
>>>> wrote:
>>>>
>>>> yes, right. What you can do is edit the hosted-engine.conf file and
>>>> there is a parameter as shown below [1] and replace h2 and h3 with your
>>>> second and third storage servers. Then you will need to restart
>>>> ovirt-ha-agent and ovirt-ha-broker services in all the nodes .
>>>>
>>>> [1] 'mnt_options=backup-volfile-servers=<h2>:<h3>'
>>>>
>>>> On Thu, Aug 31, 2017 at 5:54 PM, Charles Kozler
<ckozleriii(a)gmail.com>
>>>> wrote:
>>>>
>>>> Hi Kasturi -
>>>>
>>>> Thanks for feedback
>>>>
>>>> > If cockpit+gdeploy plugin would be have been used then that would
>>>> have automatically detected glusterfs replica 3 volume created during
>>>> Hosted Engine deployment and this question would not have been asked
>>>>
>>>> Actually, doing hosted-engine --deploy it too also auto detects
>>>> glusterfs. I know glusterfs fuse client has the ability to failover
>>>> between all nodes in cluster, but I am still curious given the fact that
I
>>>> see in ovirt config node1:/engine (being node1 I set it to in
hosted-engine
>>>> --deploy). So my concern was to ensure and find out exactly how engine
>>>> works when one node goes away and the fuse client moves over to the
other
>>>> node in the gluster cluster
>>>>
>>>> But you did somewhat answer my question, the answer seems to be no (as
>>>> default) and I will have to use hosted-engine.conf and change the
parameter
>>>> as you list
>>>>
>>>> So I need to do something manual to create HA for engine on gluster?
>>>> Yes?
>>>>
>>>> Thanks so much!
>>>>
>>>> On Thu, Aug 31, 2017 at 3:03 AM, Kasturi Narra <knarra(a)redhat.com>
>>>> wrote:
>>>>
>>>> Hi,
>>>>
>>>> During Hosted Engine setup question about glusterfs volume is being
>>>> asked because you have setup the volumes yourself. If cockpit+gdeploy
>>>> plugin would be have been used then that would have automatically
detected
>>>> glusterfs replica 3 volume created during Hosted Engine deployment and
this
>>>> question would not have been asked.
>>>>
>>>> During new storage domain creation when glusterfs is selected there
>>>> is a feature called 'use managed gluster volumes' and upon
checking this
>>>> all glusterfs volumes managed will be listed and you could choose the
>>>> volume of your choice from the dropdown list.
>>>>
>>>> There is a conf file called /etc/hosted-engine/hosted-engine.conf
>>>> where there is a parameter called
backup-volfile-servers="h1:h2" and if one
>>>> of the gluster node goes down engine uses this parameter to provide ha /
>>>> failover.
>>>>
>>>> Hope this helps !!
>>>>
>>>> Thanks
>>>> kasturi
>>>>
>>>>
>>>>
>>>> On Wed, Aug 30, 2017 at 8:09 PM, Charles Kozler
<ckozleriii(a)gmail.com>
>>>> wrote:
>>>>
>>>> Hello -
>>>>
>>>> I have successfully created a hyperconverged hosted engine setup
>>>> consisting of 3 nodes - 2 for VM's and the third purely for storage.
I
>>>> manually configured it all, did not use ovirt node or anything. Built
the
>>>> gluster volumes myself
>>>>
>>>> However, I noticed that when setting up the hosted engine and even
>>>> when adding a new storage domain with glusterfs type, it still asks for
>>>> hostname:/volumename
>>>>
>>>> This leads me to believe that if that one node goes down (ex:
>>>> node1:/data), then ovirt engine wont be able to communicate with that
>>>> volume because its trying to reach it on node 1 and thus, go down
>>>>
>>>> I know glusterfs fuse client can connect to all nodes to provide
>>>> failover/ha but how does the engine handle this?
>>>>
>>>> _______________________________________________
>>>> Users mailing list
>>>> Users(a)ovirt.org
>>>>
http://lists.ovirt.org/mailman/listinfo/users
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> _______________________________________________
>>>> Users mailing list
>>>> Users(a)ovirt.org
>>>>
http://lists.ovirt.org/mailman/listinfo/users
>>>>
>>>>
>>>> _______________________________________________
>>>> Users mailing
listUsers@ovirt.orghttp://lists.ovirt.org/mailman/listinfo/users
>>>>
>>>>
>>>> _______________________________________________
>>>> Users mailing list
>>>> Users(a)ovirt.org
>>>>
http://lists.ovirt.org/mailman/listinfo/users
>>>>
>>>>
>>>
>>
>> _______________________________________________
>> Users mailing list
>> Users(a)ovirt.org
>>
http://lists.ovirt.org/mailman/listinfo/users
>>
>>
>