moin,
i have learned to install a self hosted engine directly to the physical interfaces.
later you can move it with the hosted engine into the different bonds or vlans.
it works fine by me round 20 times.
br
marcel
Am 3. Februar 2021 19:56:47 MEZ schrieb Nardus Geldenhuys <nardusg(a)gmail.com>:
Hi oVirt land
Hope you are well. Running into this issue, I hope you can help.
Centos7 and it is updated.
Ovirt 4.3, latest packages.
My network config:
[root@mob-r1-d-ovirt-aa-1-01 ~]# ip a
1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN
group
default qlen 1000
link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
inet 127.0.0.1/8 scope host lo
valid_lft forever preferred_lft forever
inet6 ::1/128 scope host
valid_lft forever preferred_lft forever
2: ens1f0: <BROADCAST,MULTICAST,SLAVE,UP,LOWER_UP> mtu 1500 qdisc mq
master
bond0 state UP group default qlen 1000
link/ether 00:90:fa:c2:d2:48 brd ff:ff:ff:ff:ff:ff
3: ens1f1: <BROADCAST,MULTICAST,SLAVE,UP,LOWER_UP> mtu 1500 qdisc mq
master
bond0 state UP group default qlen 1000
link/ether 00:90:fa:c2:d2:48 brd ff:ff:ff:ff:ff:ff
4: enp11s0f0: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500 qdisc mq
state
DOWN group default qlen 1000
link/ether 00:90:fa:c2:d2:50 brd ff:ff:ff:ff:ff:ff
5: enp11s0f1: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500 qdisc mq
state
DOWN group default qlen 1000
link/ether 00:90:fa:c2:d2:54 brd ff:ff:ff:ff:ff:ff
21: bond0: <BROADCAST,MULTICAST,MASTER,UP,LOWER_UP> mtu 1500 qdisc
noqueue
state UP group default qlen 1000
link/ether 00:90:fa:c2:d2:48 brd ff:ff:ff:ff:ff:ff
inet6 fe80::290:faff:fec2:d248/64 scope link
valid_lft forever preferred_lft forever
22: bond0.1131@bond0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
noqueue state UP group default qlen 1000
link/ether 00:90:fa:c2:d2:48 brd ff:ff:ff:ff:ff:ff
inet 172.18.206.184/23 brd 172.18.207.255 scope global bond0.1131
valid_lft forever preferred_lft forever
inet6 fe80::290:faff:fec2:d248/64 scope link
valid_lft forever preferred_lft forever
[root@mob-r1-d-ovirt-aa-1-01 network-scripts]# cat ifcfg-bond0
BONDING_OPTS='mode=1 miimon=100'
TYPE=Bond
BONDING_MASTER=yes
PROXY_METHOD=none
BROWSER_ONLY=no
IPV6INIT=no
NAME=bond0
UUID=c11ef6ef-794f-4683-a068-d6338e5c19b6
DEVICE=bond0
ONBOOT=yes
[root@mob-r1-d-ovirt-aa-1-01 network-scripts]# cat ifcfg-bond0.1131
DEVICE=bond0.1131
VLAN=yes
ONBOOT=yes
MTU=1500
IPADDR=172.18.206.184
NETMASK=255.255.254.0
GATEWAY=172.18.206.1
BOOTPROTO=none
MTU=1500
DEFROUTE=yes
NM_CONTROLLED=no
IPV6INIT=yes
DNS1=172.20.150.10
DNS2=172.20.150.11
I get the following error:
[ INFO ] TASK [ovirt.hosted_engine_setup : Generate output list]
[ INFO ] ok: [localhost]
[ INFO ] TASK [ovirt.hosted_engine_setup : Validate selected bridge
interface if management bridge does not exists]
[ INFO ] skipping: [localhost]
Please indicate a nic to set ovirtmgmt bridge on: (bond0,
bond0.1131) [bond0.1131]:
Please specify which way the network connectivity should be
checked (ping, dns, tcp, none) [dns]:
..
..
..
..
..
[ INFO ] ok: [localhost]
[ INFO ] TASK [ovirt.hosted_engine_setup : Validate selected bridge
interface if management bridge does not exists]
[ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg":
"The
selected network interface is not valid"}
[ ERROR ] Failed to execute stage 'Closing up': Failed executing
ansible-playbook
[ INFO ] Stage: Clean up
And if I create the ifcfg-ovirtmgmt as a bridge it fails later.
What is the correct network setup for my bond configuration to do a
self
hosted-engine setup ?
Regards
Nar