[Users] Keepalived on oVirt Hosts has engine networking issues
Andrew Lau
andrew at andrewklau.com
Sun Dec 1 09:30:22 UTC 2013
I put the management and storage on separate VLANs to try avoid the
floating IP address issue temporarily. I also bonded the two nics, but I
don't think that shouldn't matter.
The other server got brought down the other day for some maintenance, I
hope to get it back up in a few days. But I can tell you a few things I
noticed:
ip a - it'll list the floating IP on both servers even if only active on
one.
I've got about 10 other networks so I've snipped out quite a bit.
# ip a
<snip>
130: bond0.2 at bond0: <BROADCAST,MULTICAST,MASTER,UP,LOWER_UP> mtu 1500 qdisc
noqueue state UP
link/ether 00:10:18:2e:6a:cb brd ff:ff:ff:ff:ff:ff
inet 172.16.0.11/24 brd 172.16.0.255 scope global bond0.2
inet6 fe80::210:18ff:fe2e:6acb/64 scope link
valid_lft forever preferred_lft forever
131: bond0.3 at bond0: <BROADCAST,MULTICAST,MASTER,UP,LOWER_UP> mtu 1500 qdisc
noqueue state UP
link/ether 00:10:18:2e:6a:cb brd ff:ff:ff:ff:ff:ff
inet 172.16.1.11/24 brd 172.16.1.255 scope global bond0.3
inet 172.16.1.5/32 scope global bond0.3
inet6 fe80::210:18ff:fe2e:6acb/64 scope link
valid_lft forever preferred_lft forever
</snip>
# vdsClient -s 0 getVdsCaps
<snip>
'storage_network': {'addr': '172.16.1.5',
'bridged': False,
'gateway': '172.16.1.1',
'iface': 'bond0.3',
'interface': 'bond0.3',
'ipv6addrs':
['fe80::210:18ff:fe2e:6acb/64'],
'ipv6gateway': '::',
'mtu': '1500',
'netmask': '255.255.255.255',
'qosInbound': '',
'qosOutbound': ''},
<snip>
vlans = {'bond0.2': {'addr': '172.16.0.11',
'cfg': {'BOOTPROTO': 'none',
'DEFROUTE': 'yes',
'DEVICE': 'bond0.2',
'GATEWAY': '172.16.0.1',
'IPADDR': '172.16.0.11',
'NETMASK': '255.255.255.0',
'NM_CONTROLLED': 'no',
'ONBOOT': 'yes',
'VLAN': 'yes'},
'iface': 'bond0',
'ipv6addrs': ['fe80::210:18ff:fe2e:6acb/64'],
'mtu': '1500',
'netmask': '255.255.255.0',
'vlanid': 2},
'bond0.3': {'addr': '172.16.1.5',
'cfg': {'BOOTPROTO': 'none',
'DEFROUTE': 'no',
'DEVICE': 'bond0.3',
'IPADDR': '172.16.1.11',
'NETMASK': '255.255.255.0',
'NM_CONTROLLED': 'no',
'ONBOOT': 'yes',
'VLAN': 'yes'},
'iface': 'bond0',
'ipv6addrs': ['fe80::210:18ff:fe2e:6acb/64'],
'mtu': '1500',
'netmask': '255.255.255.255',
'vlanid': 3},
I hope that's enough info, if not I'll post the full config on both when I
can bring it back up.
Cheers,
Andrew.
On Sun, Dec 1, 2013 at 7:15 PM, Assaf Muller <amuller at redhat.com> wrote:
> Could you please attach the output of:
> "vdsClient -s 0 getVdsCaps"
> (Or without the -s, whichever works)
> And:
> "ip a"
>
> On both hosts?
> You seem to have made changes since the documentation on the link you
> provided, like separating the management and storage via VLANs on eth0. Any
> other changes?
>
>
> Assaf Muller, Cloud Networking Engineer
> Red Hat
>
> ----- Original Message -----
> From: "Andrew Lau" <andrew at andrewklau.com>
> To: "users" <users at ovirt.org>
> Sent: Sunday, December 1, 2013 4:55:32 AM
> Subject: [Users] Keepalived on oVirt Hosts has engine networking issues
>
> Hi,
>
> I have the scenario where I have gluster and ovirt hosts on the same box,
> to keep the gluster volumes highly available incase a box drops I'm using
> keepalived across the boxes and using that IP as the means for the storage
> domain. I documented my setup here in case anyone needs a little more info
> http://www.andrewklau.com/returning-to-glusterized-ovirt-3-3/
>
> However, the engine seems to be picking up the floating IP assigned to
> keepalived as the interface and messing with the ovirtmgmt migration
> network, so migrations are failing as my floating IP gets assigned to the
> ovirtmgmt bridge in the engine however it's not actually there on most
> hosts (except one) so vdsm seems to report destination same as source.
>
> I've since created a new vlan interface just for storage to avoid the
> ovirtmgmt conflict, but the engine will still pick up the wrong IP on the
> storage vlan because of keepalived. This means I can't use the save network
> feature within the engine as it'll save the floating ip rather than the one
> already there. Is this a bug or just the way it's designed.
>
> eth0.2 -> ovirtmgmt (172.16.0.11) -> management and migration network ->
> engine sees, sets and saves 172.16.0.11
> eth0.3 -> storagenetwork (172.16.1.11) -> gluster network -> engine sees,
> sets and saves 172.16.1.5 (my floating IP)
>
> I hope this makes sense.
>
> p.s. can anyone also confirm, does gluster support multi pathing by
> default? If I'm using this keepalived method, am I bottle necking myself to
> one host?
>
> Thanks,
> Andrew
>
> _______________________________________________
> Users mailing list
> Users at ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20131201/8019843c/attachment-0001.html>
More information about the Users
mailing list