[ovirt-users] Upgrading HC from 4.0 to 4.1
Sahina Bose
sabose at redhat.com
Mon Jul 3 10:48:13 UTC 2017
On Sun, Jul 2, 2017 at 12:21 AM, Doug Ingham <dougti at gmail.com> wrote:
>
> Only problem I would like to manage is that I have gluster network shared
>> with ovirtmgmt one.
>> Can I move it now with these updated packages?
>>
>
> Are the gluster peers configured with the same hostnames/IPs as your hosts
> within oVirt?
>
> Once they're configured on the same network, separating them might be a
> bit difficult. Also, the last time I looked, oVirt still doesn't support
> managing HCI oVirt/Gluster nodes running each service on a different
> interface (see below).
>
> In theory, the procedure would involve stopping all of the Gluster
> processes on all of the peers, updating the peer addresses in the gluster
> configs on all of the nodes, then restarting glusterd & the bricks. I've
> not tested this however, and it's not a "supported" procedure. I've no idea
> how oVirt would deal with these changes either.
>
Which version of glusterfs do you have running now? With glusterfs>= 3.9,
there's a reset-brick command that can help you do this.
It's possible to move to the new interface for gluster.
The procedure would be:
1. Create a network with "gluster" network role.
2. On each host, use "Setup networks" to associate the gluster network on
the desired interface. (This would ensure thet the engine will peer probe
this interface's IP address as well, so that it can be used to identify the
host in brick defintion)
3. For each of the volume's bricks - change the definition of the brick, so
that the new ip address is used. Ensure that there's no pending heal (i.e
gluster volume heal info - should list 0 entires) before you start this(see
https://gluster.readthedocs.io/en/latest/release-notes/3.9.0/ - Introducing
reset-brick command)
gluster volume reset-brick VOLNAME <original-hostname/ip>:BRICKPATH start
gluster volume reset-brick VOLNAME <original-hostname/ip>:BRICKPATH
<new-gluster-interface-ip>:BRICKPATH commit force
>
>
> To properly separate my own storage & management networks from the
> beginning, I configured each host with 2 IPs on different subnets and a
> different hostname corresponding to each IP. For example, "v0" points to
> the management interface of the first node, and "s0" points to the storage
> interface.
>
> oVirt's problem is that, whilst it can see the pre-configured bricks and
> volumes on each host, it can't create any new bricks or volumes because it
> wants to use the same hostnames it uses to manage the hosts. It also means
> that it can't fence the hosts correctly, as it doesn't understand that "v0"
> & "s0" are the same host.
> This isn't a problem for me though, as I don't need to manage my Gluster
> instances via the GUI, and automatic fencing can be done via the IPMI
> interfaces.
>
> Last I read, this is a recognised problem, but a fix isn't expect to
> arrive any time soon.
>
> --
> Doug
>
> _______________________________________________
> Users mailing list
> Users at ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20170703/2502c5a4/attachment.html>
More information about the Users
mailing list