<html>
<head>
<meta content="text/html; charset=utf-8" http-equiv="Content-Type">
</head>
<body text="#000000" bgcolor="#FFFFFF">
<div class="moz-cite-prefix">On 12/19/2015 05:53 AM, Gianluca Cecchi
wrote:<br>
</div>
<blockquote
cite="mid:CAG2kNCx72GAq1CNfmisV_Ohec4v+MVdZ-UPzooyoyTi9Xbo4Zw@mail.gmail.com"
type="cite">
<div dir="ltr">
<div class="gmail_extra">
<div class="gmail_quote">On Sat, Dec 19, 2015 at 1:08 AM, John
Florian <span dir="ltr"><<a moz-do-not-send="true"
href="mailto:jflorian@doubledog.org" target="_blank">jflorian@doubledog.org</a>></span>
wrote:<br>
<blockquote class="gmail_quote" style="margin:0px 0px 0px
0.8ex;border-left:1px solid
rgb(204,204,204);padding-left:1ex">
<div text="#000000" bgcolor="#FFFFFF"> I'm trying to get a
3.6.1 HE setup going where I have 4 VLANs (VIDs 101-104)
for storage networks, 1 VLAN (VID 100) for ovirtmgmt and
1 more (VID 1) for everything else. Because I know of
no way to manipulate the network configuration from the
management GUI once the HE is running and with only a
single Host, I made the OS configuration as close as
possible to what I'd want when done. This looks like:<br>
</div>
</blockquote>
</div>
<br>
</div>
<div class="gmail_extra">Why do you think of this necessary
pre-work?</div>
</div>
</blockquote>
<br>
Because my storage is iSCSI and I need the VLAN configuration in
place for the Host to access it on behalf of the HE. Otherwise, yes
I agree it would be easier to let the hosted-engine script deal with
the set up. I've done a workable setup before letting the script do
everything, but the mode 4 bonding only gave me half the possible
performance because in effect one NIC on the NAS did all the
transmitting while the other NIC did all the receiving. So I really
need all of the storage network setup in place prior to starting the
HE deployment.<br>
<br>
It seems like it should be trivial to convince the engine that the
two netmasks are indeed equivalent. I tried changing in
/var/lib/vdsm/persistence/netconf/nets/ovirtmgmt the '"prefix":
"24"' setting to '"netmask": "255.255.255.0"' and running
/usr/share/vdsm/vdsm-restore-net-config but that didn't seem to
change anything WRT the network being out of sync.<br>
<br>
<blockquote
cite="mid:CAG2kNCx72GAq1CNfmisV_Ohec4v+MVdZ-UPzooyoyTi9Xbo4Zw@mail.gmail.com"
type="cite">
<div dir="ltr">
<div class="gmail_extra"> I configured (in 3.6.0) an environment
with HE too on a single host and I only preconfigured my bond1
in 802.3ad mode with the interfaces I planned to use for
ovirtmgmt and I left the other interfaces unconfigured, so
that all is not used by Network Manager.<br>
</div>
<div class="gmail_extra">During the "hosted-engine --deploy"
setup I got this input:<br>
<br>
--== NETWORK CONFIGURATION ==--<br>
<br>
Please indicate a nic to set ovirtmgmt bridge on:
(em1, bond1, em2) [em1]: bond1<br>
iptables was detected on your computer, do you wish
setup to configure it? (Yes, No)[Yes]: <br>
Please indicate a pingable gateway IP address
[10.4.168.254]: <br>
<br>
</div>
<div class="gmail_extra">and then on preview of configuration to
apply:<br>
<br>
--== CONFIGURATION PREVIEW ==--<br>
<br>
Bridge interface : bond1<br>
Engine FQDN :
ractorshe.mydomain.local<br>
Bridge name : ovirtmgmt<br>
<br>
</div>
<div class="gmail_extra">After setup I configured my vlan based
networks for my VMS from the GUI itself as in the usual way,
so that now I have this bond0 created by oVirt GUI on the
other two interfaces (em1 and em2):<br>
<br>
[root@ractor ~]# cat /proc/net/bonding/bond0<br>
Ethernet Channel Bonding Driver: v3.7.1 (April 27, 2011)<br>
<br>
Bonding Mode: IEEE 802.3ad Dynamic link aggregation<br>
Transmit Hash Policy: layer2 (0)<br>
MII Status: up<br>
MII Polling Interval (ms): 100<br>
Up Delay (ms): 0<br>
Down Delay (ms): 0<br>
<br>
802.3ad info<br>
LACP rate: fast<br>
Min links: 0<br>
Aggregator selection policy (ad_select): stable<br>
Active Aggregator Info:<br>
Aggregator ID: 2<br>
Number of ports: 2<br>
Actor Key: 17<br>
Partner Key: 8<br>
Partner Mac Address: 00:01:02:03:04:0c<br>
<br>
Slave Interface: em1<br>
MII Status: up<br>
Speed: 1000 Mbps<br>
Duplex: full<br>
Link Failure Count: 0<br>
Permanent HW addr: 00:25:64:ff:0b:f0<br>
Aggregator ID: 2<br>
Slave queue ID: 0<br>
<br>
Slave Interface: em2<br>
MII Status: up<br>
Speed: 1000 Mbps<br>
Duplex: full<br>
Link Failure Count: 0<br>
Permanent HW addr: 00:25:64:ff:0b:f2<br>
Aggregator ID: 2<br>
Slave queue ID: 0<br>
<br>
</div>
<div class="gmail_extra">And then "ip a" command returns:<br>
<br>
9: bond0.65@bond0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu
1500 qdisc noqueue master vlan65 state UP <br>
link/ether 00:25:64:ff:0b:f0 brd ff:ff:ff:ff:ff:ff<br>
10: vlan65: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500
qdisc noqueue state UP <br>
link/ether 00:25:64:ff:0b:f0 brd ff:ff:ff:ff:ff:ff<br>
<br>
with<br>
[root@ractor ~]# brctl show<br>
bridge name bridge id STP enabled interfaces<br>
;vdsmdummy; 8000.000000000000 no <br>
ovirtmgmt 8000.002564ff0bf4 no bond1<br>
vnet0<br>
vlan65 8000.002564ff0bf0 no bond0.65<br>
vnet1<br>
vnet2<br>
<br>
</div>
<div class="gmail_extra">vnet1 and vnet2 being the virtual
network interfaces of my two running VMs.<br>
<br>
</div>
<div class="gmail_extra">The only note I can submit is that by
default when you set a network in oVirt GUI with mode=4
(802.3ad), it defaults to configuring it with "lacp_rate=0" so
slow, that I think it is bad, as I read in many articles (but
I'm not a network guru at all)<br>
</div>
<div class="gmail_extra">So that I chose custom mode in the GUI
and specified "mode=4 lacp_rate=1" in options and this was
reflected in my configuration as you see above in bond0
output.<br>
<br>
</div>
<div class="gmail_extra">Can we set lacp_rate=1 as a default
option for mode=4 in oVirt?<br>
</div>
<div class="gmail_extra"><br>
</div>
<div class="gmail_extra">HIH,<br>
</div>
<div class="gmail_extra">Gianluca <br>
</div>
</div>
</blockquote>
<br>
<br>
<pre class="moz-signature" cols="72">--
John Florian
</pre>
</body>
</html>