Re: losing ib0 connection after activating host

THANK YOU SO MUCH! Thanks, Douglas Duckworth, MSc, LFCS HPC System Administrator Scientific Computing Unit<https://scu.med.cornell.edu> Weill Cornell Medicine 1300 York Avenue New York, NY 10065 E: doug@med.cornell.edu<mailto:doug@med.cornell.edu> O: 212-746-6305 F: 212-746-8690 On Tue, Dec 11, 2018 at 2:03 AM Dominik Holler <dholler@redhat.com<mailto:dholler@redhat.com>> wrote: On Mon, 10 Dec 2018 18:09:40 +0000 Douglas Duckworth <dod2014@med.cornell.edu<mailto:dod2014@med.cornell.edu>> wrote:
Hi Dominik,
I have added LACP bond network to all hosts and renamed the Hosted Engine using "/usr/share/ovirt-engine/setup/bin/ovirt-engine-rename."
However, I am still missing the option to assign Migration and Management network roles to this new bond.
Can you advise where I can find this option?
You cannot assign this role to the host interface directly, but to the network, which is assigned to the interface, in "Compute > Clusters > Clustername > Logical Networks > Manage Networks"
Thanks,
Douglas Duckworth, MSc, LFCS HPC System Administrator Scientific Computing Unit<https://scu.med.cornell.edu> Weill Cornell Medicine 1300 York Avenue New York, NY 10065 E: doug@med.cornell.edu<mailto:doug@med.cornell.edu><mailto:doug@med.cornell.edu<mailto:doug@med.cornell.edu>> O: 212-746-6305 F: 212-746-8690
On Fri, Aug 24, 2018 at 11:52 AM Dominik Holler <dholler@redhat.com<mailto:dholler@redhat.com><mailto:dholler@redhat.com<mailto:dholler@redhat.com>>> wrote: On Fri, 24 Aug 2018 09:46:25 -0400 Douglas Duckworth <dod2014@med.cornell.edu<mailto:dod2014@med.cornell.edu><mailto:dod2014@med.cornell.edu<mailto:dod2014@med.cornell.edu>>> wrote:
Sorry, I mean "migration network" for moving live migration traffic.
You have to create a new logical network in "Network > Networks > New" and assign this to ib0 in "Compute > Hosts > hostname > Network Interfaces > Setup Host Networks". After this you can assign a role to this network in "Compute > Clusters > Clustername > Logical Networks > Manage Networks"
FDR infiniband much faster than 1Gb network which currently acts as migration network, vm network, display network, mgmt network, etc.
Thanks,
Douglas Duckworth, MSc, LFCS HPC System Administrator Scientific Computing Unit Weill Cornell Medicine 1300 York - LC-502 E: doug@med.cornell.edu<mailto:doug@med.cornell.edu><mailto:doug@med.cornell.edu<mailto:doug@med.cornell.edu>> O: 212-746-6305 F: 212-746-8690
On Fri, Aug 24, 2018 at 9:36 AM, Dominik Holler <dholler@redhat.com<mailto:dholler@redhat.com><mailto:dholler@redhat.com<mailto:dholler@redhat.com>>> wrote:
On Thu, 23 Aug 2018 13:51:39 -0400 Douglas Duckworth <dod2014@med.cornell.edu<mailto:dod2014@med.cornell.edu><mailto:dod2014@med.cornell.edu<mailto:dod2014@med.cornell.edu>>> wrote:
THANKS!
ib0 now up with NFS storage back on this hypervisor
Thanks for letting us know.
Though how do I make it a transfer network? I don't see an option.
I do not understand the meaning of "transfer network". The network interface to use for NFS results from the routing tables of the host. In "Compute > Clusters > Clustername > Logical Networks > Manage Networks" network roles for some kind of loads can be assigned, but not for NFS access.
Thanks,
Douglas Duckworth, MSc, LFCS HPC System Administrator Scientific Computing Unit Weill Cornell Medicine 1300 York - LC-502 E: doug@med.cornell.edu<mailto:doug@med.cornell.edu><mailto:doug@med.cornell.edu<mailto:doug@med.cornell.edu>> O: 212-746-6305 F: 212-746-8690
On Thu, Aug 23, 2018 at 11:12 AM, Douglas Duckworth <dod2014@med.cornell.edu<mailto:dod2014@med.cornell.edu><mailto:dod2014@med.cornell.edu<mailto:dod2014@med.cornell.edu>>
wrote:
Hi Dominik
Yes, the network-script was created by our Ansible role that deploys CentOS hosts. It pulls the IP from DNS then templates the script and copies to host.
I will try this oVirt step then see if it works!
Thanks,
Douglas Duckworth, MSc, LFCS HPC System Administrator Scientific Computing Unit Weill Cornell Medicine 1300 York - LC-502 E: doug@med.cornell.edu<mailto:doug@med.cornell.edu><mailto:doug@med.cornell.edu<mailto:doug@med.cornell.edu>> O: 212-746-6305 F: 212-746-8690
On Thu, Aug 23, 2018 at 11:09 AM, Dominik Holler <dholler@redhat.com<mailto:dholler@redhat.com><mailto:dholler@redhat.com<mailto:dholler@redhat.com>>> wrote:
Is ifcfg-ib0 created before adding the host? Can ib0 be reconfigured using engine, e.g. by "Compute > Hosts > hostx > Network Interfaces > Setup Host Networks"? If this some kind of self-hosted engine?
On Thu, 23 Aug 2018 09:30:59 -0400 Douglas Duckworth <dod2014@med.cornell.edu<mailto:dod2014@med.cornell.edu><mailto:dod2014@med.cornell.edu<mailto:dod2014@med.cornell.edu>>> wrote:
> Here's a link to the files: > > https://urldefense.proofpoint.com/v2/url?u=https-3A__bit.ly_ 2wjZ6Vo&d=DwICAg&c=lb62iw4YL4RFalcE2hQUQealT9-RXrryqt9KZX2qu 2s&r=2Fzhh_78OGspKQpl_e-CbhH6xUjnRkaqPFUS2wTJ2cw&m=Y25- OOvgu58jlC82-fzBeNIpQ7ZscoHznffUhqE6EBM&s=QQXlC9Tisa60TvimyS 3BnFDCaDF7VPD8eCzT-Fke-p0&e= > > Thank you! > > Thanks, > > Douglas Duckworth, MSc, LFCS > HPC System Administrator > Scientific Computing Unit > Weill Cornell Medicine > 1300 York - LC-502 > E: doug@med.cornell.edu<mailto:doug@med.cornell.edu><mailto:doug@med.cornell.edu<mailto:doug@med.cornell.edu>> > O: 212-746-6305 > F: 212-746-8690 > > > On Thu, Aug 23, 2018 at 6:51 AM, Dominik Holler > <dholler@redhat.com<mailto:dholler@redhat.com><mailto:dholler@redhat.com<mailto:dholler@redhat.com>>> wrote: > > > Would you please share the vdsm.log and the supervdsm.log > > from this host? > > > > On Wed, 22 Aug 2018 11:36:09 -0400 > > Douglas Duckworth <dod2014@med.cornell.edu<mailto:dod2014@med.cornell.edu><mailto:dod2014@med.cornell.edu<mailto:dod2014@med.cornell.edu>>> wrote: > > > > > Hi > > > > > > I keep losing ib0 connection on hypervisor after adding > > > host to engine. This makes the host not really work > > > since NFS will be mounted over ib0. > > > > > > I don't really understand why this occurs. > > > > > > OS: > > > > > > [root@ovirt-hv2 ~]# cat /etc/redhat-release > > > CentOS Linux release 7.5.1804 (Core) > > > > > > Here's the network script: > > > > > > [root@ovirt-hv2 ~]# > > > cat /etc/sysconfig/network-scripts/ifcfg-ib0 DEVICE=ib0 > > > BOOTPROTO=static > > > IPADDR=172.16.0.207 > > > NETMASK=255.255.255.0 > > > ONBOOT=yes > > > ZONE=public > > > > > > When I try "ifup" > > > > > > [root@ovirt-hv2 ~]# ifup ib0 > > > Error: Connection activation failed: No suitable device > > > found for this connection. > > > > > > The error in syslog: > > > > > > Aug 22 11:31:50 ovirt-hv2 kernel: IPv4: martian source > > > 172.16.0.87 from 172.16.0.49, on dev ib0 > > > Aug 22 11:31:53 ovirt-hv2 NetworkManager[1070]: <info> > > > [1534951913.7486] audit: op="connection-activate" > > > uuid="2ab4abde-b8a5-6cbc-19b1-2bfb193e4e89" name="System > > > ib0" result="fail" reason="No suitable device found for > > > this connection. > > > > > > As you can see media state up: > > > > > > [root@ovirt-hv2 ~]# ip a > > > 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue > > > state UNKNOWN group default qlen 1000 > > > link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 > > > inet 127.0.0.1/8<http://127.0.0.1/8><https://urldefense.proofpoint.com/v2/url?u=http-3A__127.0.0.1_8&d=DwICAg&c=lb62iw4YL4RFalcE2hQUQealT9-RXrryqt9KZX2qu2s&r=2Fzhh_78OGspKQpl_e-CbhH6xUjnRkaqPFUS2wTJ2cw&m=aXIugCJrTkuWLQkorkpswsTiNexNw54qzJy4G8yPK2Q&s=zzg2pWZVYEzhCOg-5_w5xK-N_CzLyOsKmSdEl0HKtXQ&e=> scope host lo > > > valid_lft forever preferred_lft forever > > > 2: em1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc > > > mq master ovirtmgmt state UP group default qlen 1000 > > > link/ether 50:9a:4c:89:d3:81 brd ff:ff:ff:ff:ff:ff > > > 3: em2: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500 > > > qdisc mq state DOWN group default qlen 1000 > > > link/ether 50:9a:4c:89:d3:82 brd ff:ff:ff:ff:ff:ff > > > 4: p1p1: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500 > > > qdisc mq state DOWN group default qlen 1000 > > > link/ether b4:96:91:13:ea:68 brd ff:ff:ff:ff:ff:ff > > > 5: p1p2: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500 > > > qdisc mq state DOWN group default qlen 1000 > > > link/ether b4:96:91:13:ea:6a brd ff:ff:ff:ff:ff:ff > > > 6: idrac: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 > > > qdisc pfifo_fast state UNKNOWN group default qlen 1000 > > > link/ether 50:9a:4c:89:d3:84 brd ff:ff:ff:ff:ff:ff > > > inet 169.254.0.2/16<http://169.254.0.2/16><https://urldefense.proofpoint.com/v2/url?u=http-3A__169.254.0.2_16&d=DwICAg&c=lb62iw4YL4RFalcE2hQUQealT9-RXrryqt9KZX2qu2s&r=2Fzhh_78OGspKQpl_e-CbhH6xUjnRkaqPFUS2wTJ2cw&m=aXIugCJrTkuWLQkorkpswsTiNexNw54qzJy4G8yPK2Q&s=E-_h_YPJh5AaoKQqm9_jAG1ExP37yesF6fX0FX0WjPs&e=> brd 169.254.255.255 scope global > > > idrac valid_lft forever preferred_lft forever > > > 7: ib0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 2044 qdisc > > > mq state UP group default qlen 256 > > > link/infiniband > > > a0:00:02:08:fe:80:00:00:00:00:00:00:ec:0d:9a:03:00:1d:13:41 > > > brd > > > 00:ff:ff:ff:ff:12:40:1b:ff:ff:00:00:00:00:00:00:ff:ff:ff:ff > > > 8: ovs-system: <BROADCAST,MULTICAST> mtu 1500 qdisc noop > > > state DOWN group default qlen 1000 link/ether > > > 12:b4:30:22:39:5b brd ff:ff:ff:ff:ff:ff 9: br-int: > > > <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN > > > group default qlen 1000 link/ether 3e:32:e6:66:98:49 brd > > > ff:ff:ff:ff:ff:ff 25: ovirtmgmt: > > > <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue > > > state UP group default qlen 1000 link/ether > > > 50:9a:4c:89:d3:81 brd ff:ff:ff:ff:ff:ff inet > > > 10.0.0.183/16<http://10.0.0.183/16><https://urldefense.proofpoint.com/v2/url?u=http-3A__10.0.0.183_16&d=DwICAg&c=lb62iw4YL4RFalcE2hQUQealT9-RXrryqt9KZX2qu2s&r=2Fzhh_78OGspKQpl_e-CbhH6xUjnRkaqPFUS2wTJ2cw&m=aXIugCJrTkuWLQkorkpswsTiNexNw54qzJy4G8yPK2Q&s=ICy0qanp1MpWRKMxW9Qvfh8i4V3DRj5NgKNala8AhlA&e=> brd 10.0.255.255 scope global ovirtmgmt > > > valid_lft forever preferred_lft forever 26: > > > genev_sys_6081: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu > > > 65000 qdisc noqueue master ovs-system state UNKNOWN > > > group default qlen 1000 link/ether aa:32:82:1b:01:d9 brd > > > ff:ff:ff:ff:ff:ff 27: ;vdsmdummy;: <BROADCAST,MULTICAST> > > > mtu 1500 qdisc noop state DOWN group default qlen 1000 > > > link/ether 32:ff:5d:b8:c2:b4 brd ff:ff:ff:ff:ff:ff > > > > > > The card is FDR: > > > > > > [root@ovirt-hv2 ~]# lspci -v | grep Mellanox > > > 01:00.0 Network controller: Mellanox Technologies MT27500 > > > Family [ConnectX-3] > > > Subsystem: Mellanox Technologies Device 0051 > > > > > > Latest OFED driver: > > > > > > [root@ovirt-hv2 ~]# /etc/init.d/openibd status > > > > > > HCA driver loaded > > > > > > Configured IPoIB devices: > > > ib0 > > > > > > Currently active IPoIB devices: > > > ib0 > > > Configured Mellanox EN devices: > > > > > > Currently active Mellanox devices: > > > ib0 > > > > > > The following OFED modules are loaded: > > > > > > rdma_ucm > > > rdma_cm > > > ib_ipoib > > > mlx4_core > > > mlx4_ib > > > mlx4_en > > > mlx5_core > > > mlx5_ib > > > ib_uverbs > > > ib_umad > > > ib_ucm > > > ib_cm > > > ib_core > > > mlxfw > > > mlx5_fpga_tools > > > > > > I can add an IP to ib0 using "ip addr" though I need > > > Network Manager to work with ib0. > > > > > > > > > Thanks, > > > > > > Douglas Duckworth, MSc, LFCS > > > HPC System Administrator > > > Scientific Computing Unit > > > Weill Cornell Medicine > > > 1300 York - LC-502 > > > E: doug@med.cornell.edu<mailto:doug@med.cornell.edu><mailto:doug@med.cornell.edu<mailto:doug@med.cornell.edu>> > > > O: 212-746-6305 > > > F: 212-746-8690 > > > >
participants (1)
-
Douglas Duckworth