[ovirt-users] Adding Infiniband VM Network Fails

Clint Boggio clint at theboggios.com
Wed Nov 16 20:01:17 UTC 2016


That is correct. The ib0 in all of the HV nodes are accessing iSCSI and NFS over that IB link successfully.

What we are trying to do now is create a network that utilizes the second IB port (ib1) on the cards for some of the virtual machines that live inside the environment.

> On Nov 16, 2016, at 1:40 PM, Markus Stockhausen <stockhausen at collogia.de> wrote:
> 
> Hi,
> 
> we are running Infiniband on the NFS storage network only. Did I get
> it aight that this works or do you already have issues there?
> 
> Best regards.
> 
> Markus
> 
> Web: www.collogia.de
> 
> ________________________________________
> Von: users-bounces at ovirt.org [users-bounces at ovirt.org]&quot; im Auftrag von &quot;clint at theboggios.com [clint at theboggios.com]
> Gesendet: Mittwoch, 16. November 2016 20:10
> An: users at ovirt.org
> Betreff: [ovirt-users] Adding Infiniband VM Network Fails
> 
> Good Day;
> 
> I am trying to add an infiniband VM network to the hosts on my ovirt
> deployment, and the network configuration on the hosts fails to save.
> The network bridge is added successfully, but applying the bridge to the
> IB1 nic fails with little information other than it failed.
> 
> My system:
> 
> 6 HV nodes running CentOS 7 and OV version 4
> 1 Dedicated engine running CentOS 7 and engine version 4 in 3.6 mode.
> 
> The HV nodes all have Mellanox IB cards, dual port. Port 0 is for iSCSI
> and NFS connectivity and runs fine. Port 1 is for VM usage of the 10Gb
> network.
> 
> Have any of you had any dealings with this ?
> 
> 
> _______________________________________________
> Users mailing list
> Users at ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
> =
> <InterScan_Disclaimer.txt>




More information about the Users mailing list