Hi Dominik,

I have added LACP bond network to all hosts and renamed the Hosted Engine using "/usr/share/ovirt-engine/setup/bin/ovirt-engine-rename."

However, I am still missing the option to assign Migration and Management network roles to this new bond.  

Can you advise where I can find this option?

Thanks,

Douglas Duckworth, MSc, LFCS
HPC System Administrator
Scientific Computing Unit
Weill Cornell Medicine
1300 York Avenue
New York, NY 10065
E: doug@med.cornell.edu
O: 212-746-6305
F: 212-746-8690


On Fri, Aug 24, 2018 at 11:52 AM Dominik Holler <dholler@redhat.com> wrote:
On Fri, 24 Aug 2018 09:46:25 -0400
Douglas Duckworth <dod2014@med.cornell.edu> wrote:

> Sorry, I mean "migration network" for moving live migration traffic.
>

You have to create a new logical network in
"Network > Networks > New"
and assign this to ib0 in
"Compute > Hosts > hostname > Network Interfaces > Setup Host Networks".
After this you can assign a role to this network in
"Compute > Clusters > Clustername > Logical Networks > Manage Networks"


> FDR infiniband much faster than 1Gb network which currently acts as
> migration network, vm network, display network, mgmt network, etc.
>
> Thanks,
>
> Douglas Duckworth, MSc, LFCS
> HPC System Administrator
> Scientific Computing Unit
> Weill Cornell Medicine
> 1300 York - LC-502
> E: doug@med.cornell.edu
> O: 212-746-6305
> F: 212-746-8690
>
>
> On Fri, Aug 24, 2018 at 9:36 AM, Dominik Holler <dholler@redhat.com>
> wrote:
>
> > On Thu, 23 Aug 2018 13:51:39 -0400
> > Douglas Duckworth <dod2014@med.cornell.edu> wrote:
> > 
> > > THANKS!
> > >
> > > ib0 now up with NFS storage back on this hypervisor
> > > 
> >
> > Thanks for letting us know.
> > 
> > > Though how do I make it a transfer network?  I don't see an
> > > option.
> >
> > I do not understand the meaning of "transfer network".
> > The network interface to use for NFS results from the routing
> > tables of the host.
> > In "Compute > Clusters > Clustername > Logical Networks > Manage
> > Networks" network roles for some kind of loads can be assigned, but
> > not for NFS access.
> >
> > 
> > > Thanks,
> > >
> > > Douglas Duckworth, MSc, LFCS
> > > HPC System Administrator
> > > Scientific Computing Unit
> > > Weill Cornell Medicine
> > > 1300 York - LC-502
> > > E: doug@med.cornell.edu
> > > O: 212-746-6305
> > > F: 212-746-8690
> > >
> > >
> > > On Thu, Aug 23, 2018 at 11:12 AM, Douglas Duckworth
> > > <dod2014@med.cornell.edu 
> > > > wrote: 
> > > 
> > > > Hi Dominik
> > > >
> > > > Yes, the network-script was created by our Ansible role that
> > > > deploys CentOS hosts.  It pulls the IP from DNS then templates
> > > > the script and copies to host.
> > > >
> > > > I will try this oVirt step then see if it works!
> > > >
> > > > Thanks,
> > > >
> > > > Douglas Duckworth, MSc, LFCS
> > > > HPC System Administrator
> > > > Scientific Computing Unit
> > > > Weill Cornell Medicine
> > > > 1300 York - LC-502
> > > > E: doug@med.cornell.edu
> > > > O: 212-746-6305
> > > > F: 212-746-8690
> > > >
> > > >
> > > > On Thu, Aug 23, 2018 at 11:09 AM, Dominik Holler
> > > > <dholler@redhat.com> wrote:
> > > > 
> > > >> Is ifcfg-ib0 created before adding the host?
> > > >> Can ib0 be reconfigured using engine, e.g. by
> > > >> "Compute > Hosts > hostx > Network Interfaces > Setup Host
> > > >> Networks"? If this some kind of self-hosted engine?
> > > >>
> > > >> On Thu, 23 Aug 2018 09:30:59 -0400
> > > >> Douglas Duckworth <dod2014@med.cornell.edu> wrote:
> > > >> 
> > > >> > Here's a link to the files:
> > > >> >
> > > >> > https://urldefense.proofpoint.com/v2/url?u=https-3A__bit.ly_ 
> > > >> 2wjZ6Vo&d=DwICAg&c=lb62iw4YL4RFalcE2hQUQealT9-RXrryqt9KZX2qu
> > > >> 2s&r=2Fzhh_78OGspKQpl_e-CbhH6xUjnRkaqPFUS2wTJ2cw&m=Y25-
> > > >> OOvgu58jlC82-fzBeNIpQ7ZscoHznffUhqE6EBM&s=QQXlC9Tisa60TvimyS
> > > >> 3BnFDCaDF7VPD8eCzT-Fke-p0&e= 
> > > >> >
> > > >> > Thank you!
> > > >> >
> > > >> > Thanks,
> > > >> >
> > > >> > Douglas Duckworth, MSc, LFCS
> > > >> > HPC System Administrator
> > > >> > Scientific Computing Unit
> > > >> > Weill Cornell Medicine
> > > >> > 1300 York - LC-502
> > > >> > E: doug@med.cornell.edu
> > > >> > O: 212-746-6305
> > > >> > F: 212-746-8690
> > > >> >
> > > >> >
> > > >> > On Thu, Aug 23, 2018 at 6:51 AM, Dominik Holler
> > > >> > <dholler@redhat.com> wrote:
> > > >> > 
> > > >> > > Would you please share the vdsm.log and the supervdsm.log
> > > >> > > from this host?
> > > >> > >
> > > >> > > On Wed, 22 Aug 2018 11:36:09 -0400
> > > >> > > Douglas Duckworth <dod2014@med.cornell.edu> wrote:
> > > >> > > 
> > > >> > > > Hi
> > > >> > > >
> > > >> > > > I keep losing ib0 connection on hypervisor after adding
> > > >> > > > host to engine. This makes the host not really work
> > > >> > > > since NFS will be mounted over ib0.
> > > >> > > >
> > > >> > > > I don't really understand why this occurs.
> > > >> > > >
> > > >> > > > OS:
> > > >> > > >
> > > >> > > > [root@ovirt-hv2 ~]# cat /etc/redhat-release
> > > >> > > > CentOS Linux release 7.5.1804 (Core)
> > > >> > > >
> > > >> > > > Here's the network script:
> > > >> > > >
> > > >> > > > [root@ovirt-hv2 ~]#
> > > >> > > > cat /etc/sysconfig/network-scripts/ifcfg-ib0 DEVICE=ib0
> > > >> > > > BOOTPROTO=static
> > > >> > > > IPADDR=172.16.0.207
> > > >> > > > NETMASK=255.255.255.0
> > > >> > > > ONBOOT=yes
> > > >> > > > ZONE=public
> > > >> > > >
> > > >> > > > When I try "ifup"
> > > >> > > >
> > > >> > > > [root@ovirt-hv2 ~]# ifup ib0
> > > >> > > > Error: Connection activation failed: No suitable device
> > > >> > > > found for this connection.
> > > >> > > >
> > > >> > > > The error in syslog:
> > > >> > > >
> > > >> > > > Aug 22 11:31:50 ovirt-hv2 kernel: IPv4: martian source
> > > >> > > > 172.16.0.87 from 172.16.0.49, on dev ib0
> > > >> > > > Aug 22 11:31:53 ovirt-hv2 NetworkManager[1070]: <info>
> > > >> > > > [1534951913.7486] audit: op="connection-activate"
> > > >> > > > uuid="2ab4abde-b8a5-6cbc-19b1-2bfb193e4e89" name="System
> > > >> > > > ib0" result="fail" reason="No suitable device found for
> > > >> > > > this connection.
> > > >> > > >
> > > >> > > > As you can see media state up:
> > > >> > > >
> > > >> > > > [root@ovirt-hv2 ~]# ip a
> > > >> > > > 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue
> > > >> > > > state UNKNOWN group default qlen 1000
> > > >> > > >     link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
> > > >> > > >     inet 127.0.0.1/8 scope host lo
> > > >> > > >        valid_lft forever preferred_lft forever
> > > >> > > > 2: em1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
> > > >> > > > mq master ovirtmgmt state UP group default qlen 1000
> > > >> > > >     link/ether 50:9a:4c:89:d3:81 brd ff:ff:ff:ff:ff:ff
> > > >> > > > 3: em2: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500
> > > >> > > > qdisc mq state DOWN group default qlen 1000
> > > >> > > >     link/ether 50:9a:4c:89:d3:82 brd ff:ff:ff:ff:ff:ff
> > > >> > > > 4: p1p1: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500
> > > >> > > > qdisc mq state DOWN group default qlen 1000
> > > >> > > >     link/ether b4:96:91:13:ea:68 brd ff:ff:ff:ff:ff:ff
> > > >> > > > 5: p1p2: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500
> > > >> > > > qdisc mq state DOWN group default qlen 1000
> > > >> > > >     link/ether b4:96:91:13:ea:6a brd ff:ff:ff:ff:ff:ff
> > > >> > > > 6: idrac: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500
> > > >> > > > qdisc pfifo_fast state UNKNOWN group default qlen 1000
> > > >> > > >     link/ether 50:9a:4c:89:d3:84 brd ff:ff:ff:ff:ff:ff
> > > >> > > >     inet 169.254.0.2/16 brd 169.254.255.255 scope global
> > > >> > > > idrac valid_lft forever preferred_lft forever
> > > >> > > > 7: ib0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 2044 qdisc
> > > >> > > > mq state UP group default qlen 256
> > > >> > > >     link/infiniband
> > > >> > > > a0:00:02:08:fe:80:00:00:00:00:00:00:ec:0d:9a:03:00:1d:13:41
> > > >> > > > brd
> > > >> > > > 00:ff:ff:ff:ff:12:40:1b:ff:ff:00:00:00:00:00:00:ff:ff:ff:ff
> > > >> > > > 8: ovs-system: <BROADCAST,MULTICAST> mtu 1500 qdisc noop
> > > >> > > > state DOWN group default qlen 1000 link/ether
> > > >> > > > 12:b4:30:22:39:5b brd ff:ff:ff:ff:ff:ff 9: br-int:
> > > >> > > > <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN
> > > >> > > > group default qlen 1000 link/ether 3e:32:e6:66:98:49 brd
> > > >> > > > ff:ff:ff:ff:ff:ff 25: ovirtmgmt:
> > > >> > > > <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue
> > > >> > > > state UP group default qlen 1000 link/ether
> > > >> > > > 50:9a:4c:89:d3:81 brd ff:ff:ff:ff:ff:ff inet
> > > >> > > > 10.0.0.183/16 brd 10.0.255.255 scope global ovirtmgmt
> > > >> > > > valid_lft forever preferred_lft forever 26:
> > > >> > > > genev_sys_6081: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu
> > > >> > > > 65000 qdisc noqueue master ovs-system state UNKNOWN
> > > >> > > > group default qlen 1000 link/ether aa:32:82:1b:01:d9 brd
> > > >> > > > ff:ff:ff:ff:ff:ff 27: ;vdsmdummy;: <BROADCAST,MULTICAST>
> > > >> > > > mtu 1500 qdisc noop state DOWN group default qlen 1000
> > > >> > > > link/ether 32:ff:5d:b8:c2:b4 brd ff:ff:ff:ff:ff:ff
> > > >> > > >
> > > >> > > > The card is FDR:
> > > >> > > >
> > > >> > > > [root@ovirt-hv2 ~]# lspci -v | grep Mellanox
> > > >> > > > 01:00.0 Network controller: Mellanox Technologies MT27500
> > > >> > > > Family [ConnectX-3]
> > > >> > > > Subsystem: Mellanox Technologies Device 0051
> > > >> > > >
> > > >> > > > Latest OFED driver:
> > > >> > > >
> > > >> > > > [root@ovirt-hv2 ~]# /etc/init.d/openibd status
> > > >> > > >
> > > >> > > >   HCA driver loaded
> > > >> > > >
> > > >> > > > Configured IPoIB devices:
> > > >> > > > ib0
> > > >> > > >
> > > >> > > > Currently active IPoIB devices:
> > > >> > > > ib0
> > > >> > > > Configured Mellanox EN devices:
> > > >> > > >
> > > >> > > > Currently active Mellanox devices:
> > > >> > > > ib0
> > > >> > > >
> > > >> > > > The following OFED modules are loaded:
> > > >> > > >
> > > >> > > >   rdma_ucm
> > > >> > > >   rdma_cm
> > > >> > > >   ib_ipoib
> > > >> > > >   mlx4_core
> > > >> > > >   mlx4_ib
> > > >> > > >   mlx4_en
> > > >> > > >   mlx5_core
> > > >> > > >   mlx5_ib
> > > >> > > >   ib_uverbs
> > > >> > > >   ib_umad
> > > >> > > >   ib_ucm
> > > >> > > >   ib_cm
> > > >> > > >   ib_core
> > > >> > > >   mlxfw
> > > >> > > >   mlx5_fpga_tools
> > > >> > > >
> > > >> > > > I can add an IP to ib0 using "ip addr" though I need
> > > >> > > > Network Manager to work with ib0.
> > > >> > > >
> > > >> > > >
> > > >> > > > Thanks,
> > > >> > > >
> > > >> > > > Douglas Duckworth, MSc, LFCS
> > > >> > > > HPC System Administrator
> > > >> > > > Scientific Computing Unit
> > > >> > > > Weill Cornell Medicine
> > > >> > > > 1300 York - LC-502
> > > >> > > > E: doug@med.cornell.edu
> > > >> > > > O: 212-746-6305
> > > >> > > > F: 212-746-8690 
> > > >> > >
> > > >> > > 
> > > >>
> > > >> 
> > > > 
> >
> >