THANK YOU SO MUCH!

Thanks,

Douglas Duckworth, MSc, LFCS
HPC System Administrator
Scientific Computing Unit
Weill Cornell Medicine
1300 York Avenue
New York, NY 10065
E: doug@med.cornell.edu
O: 212-746-6305
F: 212-746-8690


On Tue, Dec 11, 2018 at 2:03 AM Dominik Holler <dholler@redhat.com> wrote:
On Mon, 10 Dec 2018 18:09:40 +0000
Douglas Duckworth <dod2014@med.cornell.edu> wrote:

> Hi Dominik,
>
> I have added LACP bond network to all hosts and renamed the Hosted Engine using "/usr/share/ovirt-engine/setup/bin/ovirt-engine-rename."
>
> However, I am still missing the option to assign Migration and Management network roles to this new bond.
>
> Can you advise where I can find this option?
>

You cannot assign this role to the host interface directly, but to the
network, which is assigned to the interface, in
"Compute > Clusters > Clustername > Logical Networks > Manage Networks"

> Thanks,
>
> Douglas Duckworth, MSc, LFCS
> HPC System Administrator
> Scientific Computing Unit<https://scu.med.cornell.edu>
> Weill Cornell Medicine
> 1300 York Avenue
> New York, NY 10065
> E: doug@med.cornell.edu<mailto:doug@med.cornell.edu>
> O: 212-746-6305
> F: 212-746-8690
>
>
> On Fri, Aug 24, 2018 at 11:52 AM Dominik Holler <dholler@redhat.com<mailto:dholler@redhat.com>> wrote:
> On Fri, 24 Aug 2018 09:46:25 -0400
> Douglas Duckworth <dod2014@med.cornell.edu<mailto:dod2014@med.cornell.edu>> wrote:
>
> > Sorry, I mean "migration network" for moving live migration traffic.
> > 
>
> You have to create a new logical network in
> "Network > Networks > New"
> and assign this to ib0 in
> "Compute > Hosts > hostname > Network Interfaces > Setup Host Networks".
> After this you can assign a role to this network in
> "Compute > Clusters > Clustername > Logical Networks > Manage Networks"
>
>
> > FDR infiniband much faster than 1Gb network which currently acts as
> > migration network, vm network, display network, mgmt network, etc.
> >
> > Thanks,
> >
> > Douglas Duckworth, MSc, LFCS
> > HPC System Administrator
> > Scientific Computing Unit
> > Weill Cornell Medicine
> > 1300 York - LC-502
> > E: doug@med.cornell.edu<mailto:doug@med.cornell.edu>
> > O: 212-746-6305
> > F: 212-746-8690
> >
> >
> > On Fri, Aug 24, 2018 at 9:36 AM, Dominik Holler <dholler@redhat.com<mailto:dholler@redhat.com>>
> > wrote:
> > 
> > > On Thu, 23 Aug 2018 13:51:39 -0400
> > > Douglas Duckworth <dod2014@med.cornell.edu<mailto:dod2014@med.cornell.edu>> wrote:
> > > 
> > > > THANKS!
> > > >
> > > > ib0 now up with NFS storage back on this hypervisor
> > > > 
> > >
> > > Thanks for letting us know.
> > > 
> > > > Though how do I make it a transfer network?  I don't see an
> > > > option. 
> > >
> > > I do not understand the meaning of "transfer network".
> > > The network interface to use for NFS results from the routing
> > > tables of the host.
> > > In "Compute > Clusters > Clustername > Logical Networks > Manage
> > > Networks" network roles for some kind of loads can be assigned, but
> > > not for NFS access.
> > >
> > > 
> > > > Thanks,
> > > >
> > > > Douglas Duckworth, MSc, LFCS
> > > > HPC System Administrator
> > > > Scientific Computing Unit
> > > > Weill Cornell Medicine
> > > > 1300 York - LC-502
> > > > E: doug@med.cornell.edu<mailto:doug@med.cornell.edu>
> > > > O: 212-746-6305
> > > > F: 212-746-8690
> > > >
> > > >
> > > > On Thu, Aug 23, 2018 at 11:12 AM, Douglas Duckworth
> > > > <dod2014@med.cornell.edu<mailto:dod2014@med.cornell.edu
> > > > > wrote: 
> > > > 
> > > > > Hi Dominik
> > > > >
> > > > > Yes, the network-script was created by our Ansible role that
> > > > > deploys CentOS hosts.  It pulls the IP from DNS then templates
> > > > > the script and copies to host.
> > > > >
> > > > > I will try this oVirt step then see if it works!
> > > > >
> > > > > Thanks,
> > > > >
> > > > > Douglas Duckworth, MSc, LFCS
> > > > > HPC System Administrator
> > > > > Scientific Computing Unit
> > > > > Weill Cornell Medicine
> > > > > 1300 York - LC-502
> > > > > E: doug@med.cornell.edu<mailto:doug@med.cornell.edu>
> > > > > O: 212-746-6305
> > > > > F: 212-746-8690
> > > > >
> > > > >
> > > > > On Thu, Aug 23, 2018 at 11:09 AM, Dominik Holler
> > > > > <dholler@redhat.com<mailto:dholler@redhat.com>> wrote:
> > > > > 
> > > > >> Is ifcfg-ib0 created before adding the host?
> > > > >> Can ib0 be reconfigured using engine, e.g. by
> > > > >> "Compute > Hosts > hostx > Network Interfaces > Setup Host
> > > > >> Networks"? If this some kind of self-hosted engine?
> > > > >>
> > > > >> On Thu, 23 Aug 2018 09:30:59 -0400
> > > > >> Douglas Duckworth <dod2014@med.cornell.edu<mailto:dod2014@med.cornell.edu>> wrote:
> > > > >> 
> > > > >> > Here's a link to the files:
> > > > >> >
> > > > >> > https://urldefense.proofpoint.com/v2/url?u=https-3A__bit.ly_ 
> > > > >> 2wjZ6Vo&d=DwICAg&c=lb62iw4YL4RFalcE2hQUQealT9-RXrryqt9KZX2qu
> > > > >> 2s&r=2Fzhh_78OGspKQpl_e-CbhH6xUjnRkaqPFUS2wTJ2cw&m=Y25-
> > > > >> OOvgu58jlC82-fzBeNIpQ7ZscoHznffUhqE6EBM&s=QQXlC9Tisa60TvimyS
> > > > >> 3BnFDCaDF7VPD8eCzT-Fke-p0&e= 
> > > > >> >
> > > > >> > Thank you!
> > > > >> >
> > > > >> > Thanks,
> > > > >> >
> > > > >> > Douglas Duckworth, MSc, LFCS
> > > > >> > HPC System Administrator
> > > > >> > Scientific Computing Unit
> > > > >> > Weill Cornell Medicine
> > > > >> > 1300 York - LC-502
> > > > >> > E: doug@med.cornell.edu<mailto:doug@med.cornell.edu>
> > > > >> > O: 212-746-6305
> > > > >> > F: 212-746-8690
> > > > >> >
> > > > >> >
> > > > >> > On Thu, Aug 23, 2018 at 6:51 AM, Dominik Holler
> > > > >> > <dholler@redhat.com<mailto:dholler@redhat.com>> wrote:
> > > > >> > 
> > > > >> > > Would you please share the vdsm.log and the supervdsm.log
> > > > >> > > from this host?
> > > > >> > >
> > > > >> > > On Wed, 22 Aug 2018 11:36:09 -0400
> > > > >> > > Douglas Duckworth <dod2014@med.cornell.edu<mailto:dod2014@med.cornell.edu>> wrote:
> > > > >> > > 
> > > > >> > > > Hi
> > > > >> > > >
> > > > >> > > > I keep losing ib0 connection on hypervisor after adding
> > > > >> > > > host to engine. This makes the host not really work
> > > > >> > > > since NFS will be mounted over ib0.
> > > > >> > > >
> > > > >> > > > I don't really understand why this occurs.
> > > > >> > > >
> > > > >> > > > OS:
> > > > >> > > >
> > > > >> > > > [root@ovirt-hv2 ~]# cat /etc/redhat-release
> > > > >> > > > CentOS Linux release 7.5.1804 (Core)
> > > > >> > > >
> > > > >> > > > Here's the network script:
> > > > >> > > >
> > > > >> > > > [root@ovirt-hv2 ~]#
> > > > >> > > > cat /etc/sysconfig/network-scripts/ifcfg-ib0 DEVICE=ib0
> > > > >> > > > BOOTPROTO=static
> > > > >> > > > IPADDR=172.16.0.207
> > > > >> > > > NETMASK=255.255.255.0
> > > > >> > > > ONBOOT=yes
> > > > >> > > > ZONE=public
> > > > >> > > >
> > > > >> > > > When I try "ifup"
> > > > >> > > >
> > > > >> > > > [root@ovirt-hv2 ~]# ifup ib0
> > > > >> > > > Error: Connection activation failed: No suitable device
> > > > >> > > > found for this connection.
> > > > >> > > >
> > > > >> > > > The error in syslog:
> > > > >> > > >
> > > > >> > > > Aug 22 11:31:50 ovirt-hv2 kernel: IPv4: martian source
> > > > >> > > > 172.16.0.87 from 172.16.0.49, on dev ib0
> > > > >> > > > Aug 22 11:31:53 ovirt-hv2 NetworkManager[1070]: <info>
> > > > >> > > > [1534951913.7486] audit: op="connection-activate"
> > > > >> > > > uuid="2ab4abde-b8a5-6cbc-19b1-2bfb193e4e89" name="System
> > > > >> > > > ib0" result="fail" reason="No suitable device found for
> > > > >> > > > this connection.
> > > > >> > > >
> > > > >> > > > As you can see media state up:
> > > > >> > > >
> > > > >> > > > [root@ovirt-hv2 ~]# ip a
> > > > >> > > > 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue
> > > > >> > > > state UNKNOWN group default qlen 1000
> > > > >> > > >     link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
> > > > >> > > >     inet 127.0.0.1/8<https://urldefense.proofpoint.com/v2/url?u=http-3A__127.0.0.1_8&d=DwICAg&c=lb62iw4YL4RFalcE2hQUQealT9-RXrryqt9KZX2qu2s&r=2Fzhh_78OGspKQpl_e-CbhH6xUjnRkaqPFUS2wTJ2cw&m=aXIugCJrTkuWLQkorkpswsTiNexNw54qzJy4G8yPK2Q&s=zzg2pWZVYEzhCOg-5_w5xK-N_CzLyOsKmSdEl0HKtXQ&e=> scope host lo
> > > > >> > > >        valid_lft forever preferred_lft forever
> > > > >> > > > 2: em1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
> > > > >> > > > mq master ovirtmgmt state UP group default qlen 1000
> > > > >> > > >     link/ether 50:9a:4c:89:d3:81 brd ff:ff:ff:ff:ff:ff
> > > > >> > > > 3: em2: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500
> > > > >> > > > qdisc mq state DOWN group default qlen 1000
> > > > >> > > >     link/ether 50:9a:4c:89:d3:82 brd ff:ff:ff:ff:ff:ff
> > > > >> > > > 4: p1p1: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500
> > > > >> > > > qdisc mq state DOWN group default qlen 1000
> > > > >> > > >     link/ether b4:96:91:13:ea:68 brd ff:ff:ff:ff:ff:ff
> > > > >> > > > 5: p1p2: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500
> > > > >> > > > qdisc mq state DOWN group default qlen 1000
> > > > >> > > >     link/ether b4:96:91:13:ea:6a brd ff:ff:ff:ff:ff:ff
> > > > >> > > > 6: idrac: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500
> > > > >> > > > qdisc pfifo_fast state UNKNOWN group default qlen 1000
> > > > >> > > >     link/ether 50:9a:4c:89:d3:84 brd ff:ff:ff:ff:ff:ff
> > > > >> > > >     inet 169.254.0.2/16<https://urldefense.proofpoint.com/v2/url?u=http-3A__169.254.0.2_16&d=DwICAg&c=lb62iw4YL4RFalcE2hQUQealT9-RXrryqt9KZX2qu2s&r=2Fzhh_78OGspKQpl_e-CbhH6xUjnRkaqPFUS2wTJ2cw&m=aXIugCJrTkuWLQkorkpswsTiNexNw54qzJy4G8yPK2Q&s=E-_h_YPJh5AaoKQqm9_jAG1ExP37yesF6fX0FX0WjPs&e=> brd 169.254.255.255 scope global
> > > > >> > > > idrac valid_lft forever preferred_lft forever
> > > > >> > > > 7: ib0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 2044 qdisc
> > > > >> > > > mq state UP group default qlen 256
> > > > >> > > >     link/infiniband
> > > > >> > > > a0:00:02:08:fe:80:00:00:00:00:00:00:ec:0d:9a:03:00:1d:13:41
> > > > >> > > > brd
> > > > >> > > > 00:ff:ff:ff:ff:12:40:1b:ff:ff:00:00:00:00:00:00:ff:ff:ff:ff
> > > > >> > > > 8: ovs-system: <BROADCAST,MULTICAST> mtu 1500 qdisc noop
> > > > >> > > > state DOWN group default qlen 1000 link/ether
> > > > >> > > > 12:b4:30:22:39:5b brd ff:ff:ff:ff:ff:ff 9: br-int:
> > > > >> > > > <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN
> > > > >> > > > group default qlen 1000 link/ether 3e:32:e6:66:98:49 brd
> > > > >> > > > ff:ff:ff:ff:ff:ff 25: ovirtmgmt:
> > > > >> > > > <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue
> > > > >> > > > state UP group default qlen 1000 link/ether
> > > > >> > > > 50:9a:4c:89:d3:81 brd ff:ff:ff:ff:ff:ff inet
> > > > >> > > > 10.0.0.183/16<https://urldefense.proofpoint.com/v2/url?u=http-3A__10.0.0.183_16&d=DwICAg&c=lb62iw4YL4RFalcE2hQUQealT9-RXrryqt9KZX2qu2s&r=2Fzhh_78OGspKQpl_e-CbhH6xUjnRkaqPFUS2wTJ2cw&m=aXIugCJrTkuWLQkorkpswsTiNexNw54qzJy4G8yPK2Q&s=ICy0qanp1MpWRKMxW9Qvfh8i4V3DRj5NgKNala8AhlA&e=> brd 10.0.255.255 scope global ovirtmgmt
> > > > >> > > > valid_lft forever preferred_lft forever 26:
> > > > >> > > > genev_sys_6081: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu
> > > > >> > > > 65000 qdisc noqueue master ovs-system state UNKNOWN
> > > > >> > > > group default qlen 1000 link/ether aa:32:82:1b:01:d9 brd
> > > > >> > > > ff:ff:ff:ff:ff:ff 27: ;vdsmdummy;: <BROADCAST,MULTICAST>
> > > > >> > > > mtu 1500 qdisc noop state DOWN group default qlen 1000
> > > > >> > > > link/ether 32:ff:5d:b8:c2:b4 brd ff:ff:ff:ff:ff:ff
> > > > >> > > >
> > > > >> > > > The card is FDR:
> > > > >> > > >
> > > > >> > > > [root@ovirt-hv2 ~]# lspci -v | grep Mellanox
> > > > >> > > > 01:00.0 Network controller: Mellanox Technologies MT27500
> > > > >> > > > Family [ConnectX-3]
> > > > >> > > > Subsystem: Mellanox Technologies Device 0051
> > > > >> > > >
> > > > >> > > > Latest OFED driver:
> > > > >> > > >
> > > > >> > > > [root@ovirt-hv2 ~]# /etc/init.d/openibd status
> > > > >> > > >
> > > > >> > > >   HCA driver loaded
> > > > >> > > >
> > > > >> > > > Configured IPoIB devices:
> > > > >> > > > ib0
> > > > >> > > >
> > > > >> > > > Currently active IPoIB devices:
> > > > >> > > > ib0
> > > > >> > > > Configured Mellanox EN devices:
> > > > >> > > >
> > > > >> > > > Currently active Mellanox devices:
> > > > >> > > > ib0
> > > > >> > > >
> > > > >> > > > The following OFED modules are loaded:
> > > > >> > > >
> > > > >> > > >   rdma_ucm
> > > > >> > > >   rdma_cm
> > > > >> > > >   ib_ipoib
> > > > >> > > >   mlx4_core
> > > > >> > > >   mlx4_ib
> > > > >> > > >   mlx4_en
> > > > >> > > >   mlx5_core
> > > > >> > > >   mlx5_ib
> > > > >> > > >   ib_uverbs
> > > > >> > > >   ib_umad
> > > > >> > > >   ib_ucm
> > > > >> > > >   ib_cm
> > > > >> > > >   ib_core
> > > > >> > > >   mlxfw
> > > > >> > > >   mlx5_fpga_tools
> > > > >> > > >
> > > > >> > > > I can add an IP to ib0 using "ip addr" though I need
> > > > >> > > > Network Manager to work with ib0.
> > > > >> > > >
> > > > >> > > >
> > > > >> > > > Thanks,
> > > > >> > > >
> > > > >> > > > Douglas Duckworth, MSc, LFCS
> > > > >> > > > HPC System Administrator
> > > > >> > > > Scientific Computing Unit
> > > > >> > > > Weill Cornell Medicine
> > > > >> > > > 1300 York - LC-502
> > > > >> > > > E: doug@med.cornell.edu<mailto:doug@med.cornell.edu>
> > > > >> > > > O: 212-746-6305
> > > > >> > > > F: 212-746-8690 
> > > > >> > >
> > > > >> > > 
> > > > >>
> > > > >> 
> > > > > 
> > >
> > > 
>