
On Wed, Nov 16, 2016 at 10:01 PM, Clint Boggio <clint@theboggios.com> wrote:
That is correct. The ib0 in all of the HV nodes are accessing iSCSI and NFS over that IB link successfully.
What we are trying to do now is create a network that utilizes the second IB port (ib1) on the cards for some of the virtual machines that live inside the environment.
On Nov 16, 2016, at 1:40 PM, Markus Stockhausen <stockhausen@collogia.de> wrote:
Hi,
we are running Infiniband on the NFS storage network only. Did I get it aight that this works or do you already have issues there?
Best regards.
Markus
Web: www.collogia.de
________________________________________ Von: users-bounces@ovirt.org [users-bounces@ovirt.org]" im Auftrag von "clint@theboggios.com [clint@theboggios.com] Gesendet: Mittwoch, 16. November 2016 20:10 An: users@ovirt.org Betreff: [ovirt-users] Adding Infiniband VM Network Fails
Good Day;
I am trying to add an infiniband VM network to the hosts on my ovirt deployment, and the network configuration on the hosts fails to save. The network bridge is added successfully, but applying the bridge to the IB1 nic fails with little information other than it failed.
My system:
6 HV nodes running CentOS 7 and OV version 4 1 Dedicated engine running CentOS 7 and engine version 4 in 3.6 mode.
The HV nodes all have Mellanox IB cards, dual port. Port 0 is for iSCSI and NFS connectivity and runs fine. Port 1 is for VM usage of the 10Gb network.
Have any of you had any dealings with this ?
Could you please share Engine and node vdsm logs? (On the node, look for the vdsm.log and supervdsm.log) Thanks, Edy.
_______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users = <InterScan_Disclaimer.txt>
_______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users