Hi,
As for me personally I have such a config: compute nodes with 4x1G nics
and storages with 2x1G nics and 2 switches (not stackable). All servers
runs on CentOS 7.X (7.3 at this monent).
On compute nodes I have bonding with two nic1 and nic2 (attached to
different switches) for mgmt and VM's network, and the other two nics
nic3 and nic4 without bonding (and also attached to different switches).
On storage nodes I have no bonding and nics nic1 and nic2 connected to
different switches.
I have two networks for iSCSI: 10.0.2.0/24 and 10.0.3.0/24, nic1 of
storage and nic3 of computes connected to one network; nic2 of storage
and nic4 of computes - to another one.
On webUI I've created network iSCSI1 and iSCSI2 for nic3 and nic4, also
created multipath. To have active/active links with double bw throughput
I've added 'path_grouping_policy "multibus"' in defaults section of
/etc/multipath.conf.
After all of that, I have 200+MB/sec throughput to the storage (like
raid0 with 2 sata hdd) and I can lose one nic/link/swith without
stopping vms.
[root@compute02 ~]# multipath -ll
360014052f28c9a600000000000000000 dm-6 LIO-ORG ,ClusterLunHDD
size=902G features='0' hwhandler='0' wp=rw
`-+- policy='service-time 0' prio=1 status=active
|- 6:0:0:0 sdc 8:32 active ready running
`- 8:0:0:0 sdf 8:80 active ready running
36001405551a9610d09b4ff9aa836b906 dm-40 LIO-ORG ,SSD_DOMAIN
size=915G features='0' hwhandler='0' wp=rw
`-+- policy='service-time 0' prio=1 status=active
|- 7:0:0:0 sde 8:64 active ready running
`- 9:0:0:0 sdh 8:112 active ready running
360014055eb8d30a91044649bda9ee620 dm-5 LIO-ORG ,ClusterLunSSD
size=135G features='0' hwhandler='0' wp=rw
`-+- policy='service-time 0' prio=1 status=active
|- 6:0:0:1 sdd 8:48 active ready running
`- 8:0:0:1 sdg 8:96 active ready running
[root@compute02 ~]# iscsiadm -m session
tcp: [1] 10.0.3.200:3260,1 iqn.2015-09.lab.lnx-san:storage (non-flash)
tcp: [2] 10.0.3.203:3260,1 iqn.2016-10.local.ntu:storage3 (non-flash)
tcp: [3] 10.0.3.200:3260,1 iqn.2015-09.lab.lnx-san:storage (non-flash)
tcp: [4] 10.0.3.203:3260,1 iqn.2016-10.local.ntu:storage3 (non-flash)
[root@compute02 ~]# ip route show | head -4
default via 10.0.1.1 dev ovirtmgmt
10.0.1.0/24 dev ovirtmgmt proto kernel scope link src 10.0.1.102
10.0.2.0/24 dev enp5s0.2 proto kernel scope link src 10.0.2.102
10.0.3.0/24 dev enp2s0.3 proto kernel scope link src 10.0.3.102
[root@compute02 ~]# brctl show ovirtmgmt
bridge name bridge id STP enabled interfaces
ovirtmgmt 8000.000475b4f262 no bond0.1001
[root@compute02 ~]# cat /proc/net/bonding/bond0 | grep "Bonding\|Slave
Interface"
Ethernet Channel Bonding Driver: v3.7.1 (April 27, 2011)
Bonding Mode: fault-tolerance (active-backup)
Slave Interface: enp4s6
Slave Interface: enp6s0
01.02.2017 12:50, Nicolas Ecarnot пишет:
Hello,
I'm starting over on this subject because I wanted to clarify what was
the oVirt way to manage multipathing.
(Here I will talk only about the data/iSCSI/SAN/LUN/you name it
networks.)
According to what I see in the host network setup, one can assign
*ONE* data network to an interface or to a group of interfaces.
That implies that if my host has two data-dedicated interfaces, I can
- either group them using bonding (and oVirt is handy for that in the
host network setup), then assign the data virtual network to this bond
- either assign each nic a different ip in each a different VLAN, then
use two different data networks, and assign them each a different data
network. I never played this game and don't know where it's going to.
At first, may the oVirt storage experts comment on the above to check
it's ok.
Then, as many users here, our hardware is this :
- Hosts : Dell poweredge, mostly blades (M610,620,630), or rack servers
- SANs : Equallogic PS4xxx and PS6xxx
Equallogic's recommendation is that bonding is evil in iSCSI access.
To them, multipath is the only true way.
After reading tons of docs and using Dell support, everything is
telling me to use at least two different NICs with different ip, not
bonded - using the same network is bad but ok.
How can oVirt handle that ?