[ovirt-users] Issue with OVN/OVS and mandatory ovirtmgmt network

Sverker Abrahamsson sverker at abrahamsson.com
Thu Dec 29 20:21:12 UTC 2016


After a reboot I ran into a new problem, vdsm-network did not want to 
start. First because of that it tried to establish default route on the 
dummy interface:

Dec 29 20:57:44 h2.limetransit.com vdsm-tool[23132]: Traceback (most 
recent call last):
Dec 29 20:57:44 h2.limetransit.com vdsm-tool[23132]: File 
"/usr/share/vdsm/vdsm-restore-net-config", line 482, in <module>
Dec 29 20:57:44 h2.limetransit.com vdsm-tool[23132]: restore(args)
Dec 29 20:57:44 h2.limetransit.com vdsm-tool[23132]: File 
"/usr/share/vdsm/vdsm-restore-net-config", line 445, in restore
Dec 29 20:57:44 h2.limetransit.com vdsm-tool[23132]: unified_restoration()
Dec 29 20:57:44 h2.limetransit.com vdsm-tool[23132]: File 
"/usr/share/vdsm/vdsm-restore-net-config", line 145, in unified_restoration
Dec 29 20:57:44 h2.limetransit.com vdsm-tool[23132]: '_inRollback': True})
Dec 29 20:57:44 h2.limetransit.com vdsm-tool[23132]: File 
"/usr/lib/python2.7/site-packages/vdsm/network/api.py", line 253, in 
setupNetworks
Dec 29 20:57:44 h2.limetransit.com vdsm-tool[23132]: 
ipvalidator.validate(networks)
Dec 29 20:57:44 h2.limetransit.com vdsm-tool[23132]: File 
"/usr/lib/python2.7/site-packages/vdsm/network/ip/validator.py", line 
43, in validate
Dec 29 20:57:44 h2.limetransit.com vdsm-tool[23132]: 
_validate_default_route(default_route_nets, no_default_route_nets)
Dec 29 20:57:44 h2.limetransit.com vdsm-tool[23132]: File 
"/usr/lib/python2.7/site-packages/vdsm/network/ip/validator.py", line 
53, in _validate_default_route
Dec 29 20:57:44 h2.limetransit.com vdsm-tool[23132]: 'Only a singe 
default route network is allowed.')
Dec 29 20:57:44 h2.limetransit.com vdsm-tool[23132]: 
vdsm.network.errors.ConfigNetworkError: (21, 'Only a singe default route 
network is allowed.')
Dec 29 20:57:44 h2.limetransit.com vdsm-tool[23132]: Traceback (most 
recent call last):
Dec 29 20:57:44 h2.limetransit.com vdsm-tool[23132]: File 
"/usr/bin/vdsm-tool", line 219, in main
Dec 29 20:57:44 h2.limetransit.com vdsm-tool[23132]: return 
tool_command[cmd]["command"](*args)
Dec 29 20:57:44 h2.limetransit.com vdsm-tool[23132]: File 
"/usr/lib/python2.7/site-packages/vdsm/tool/restore_nets.py", line 41, 
in restore_command
Dec 29 20:57:44 h2.limetransit.com vdsm-tool[23132]: exec_restore(cmd)
Dec 29 20:57:44 h2.limetransit.com vdsm-tool[23132]: File 
"/usr/lib/python2.7/site-packages/vdsm/tool/restore_nets.py", line 54, 
in exec_restore
Dec 29 20:57:44 h2.limetransit.com vdsm-tool[23132]: raise 
EnvironmentError('Failed to restore the persisted networks')
Dec 29 20:57:44 h2.limetransit.com vdsm-tool[23132]: EnvironmentError: 
Failed to restore the persisted networks

I edit /var/lib/vdsm/persistence/netconf/nets/ovirtmgmt to set 
defaultroute=false, but then it complain about nameservers:

Dec 29 21:08:09 h2.limetransit.com vdsm-tool[3736]: Traceback (most 
recent call last):
Dec 29 21:08:09 h2.limetransit.com vdsm-tool[3736]: File 
"/usr/share/vdsm/vdsm-restore-net-config", line 482, in <module>
Dec 29 21:08:09 h2.limetransit.com vdsm-tool[3736]: restore(args)
Dec 29 21:08:09 h2.limetransit.com vdsm-tool[3736]: File 
"/usr/share/vdsm/vdsm-restore-net-config", line 445, in restore
Dec 29 21:08:09 h2.limetransit.com vdsm-tool[3736]: unified_restoration()
Dec 29 21:08:09 h2.limetransit.com vdsm-tool[3736]: File 
"/usr/share/vdsm/vdsm-restore-net-config", line 145, in unified_restoration
Dec 29 21:08:09 h2.limetransit.com vdsm-tool[3736]: '_inRollback': True})
Dec 29 21:08:09 h2.limetransit.com vdsm-tool[3736]: File 
"/usr/lib/python2.7/site-packages/vdsm/network/api.py", line 253, in 
setupNetworks
Dec 29 21:08:09 h2.limetransit.com vdsm-tool[3736]: 
ipvalidator.validate(networks)
Dec 29 21:08:09 h2.limetransit.com vdsm-tool[3736]: File 
"/usr/lib/python2.7/site-packages/vdsm/network/ip/validator.py", line 
36, in validate
Dec 29 21:08:09 h2.limetransit.com vdsm-tool[3736]: 
_validate_nameservers(net, attrs)
Dec 29 21:08:09 h2.limetransit.com vdsm-tool[3736]: File 
"/usr/lib/python2.7/site-packages/vdsm/network/ip/validator.py", line 
58, in _validate_nameservers
Dec 29 21:08:09 h2.limetransit.com vdsm-tool[3736]: 
_validate_nameservers_network(attrs)
Dec 29 21:08:09 h2.limetransit.com vdsm-tool[3736]: File 
"/usr/lib/python2.7/site-packages/vdsm/network/ip/validator.py", line 
66, in _validate_nameservers_network
Dec 29 21:08:09 h2.limetransit.com vdsm-tool[3736]: 'Name servers may 
only be defined on the default host network')
Dec 29 21:08:09 h2.limetransit.com vdsm-tool[3736]: 
vdsm.network.errors.ConfigNetworkError: (21, 'Name servers may only be 
defined on the default host network')
Dec 29 21:08:09 h2.limetransit.com vdsm-tool[3736]: Traceback (most 
recent call last):
Dec 29 21:08:09 h2.limetransit.com vdsm-tool[3736]: File 
"/usr/bin/vdsm-tool", line 219, in main
Dec 29 21:08:09 h2.limetransit.com vdsm-tool[3736]: return 
tool_command[cmd]["command"](*args)
Dec 29 21:08:09 h2.limetransit.com vdsm-tool[3736]: File 
"/usr/lib/python2.7/site-packages/vdsm/tool/restore_nets.py", line 41, 
in restore_command
Dec 29 21:08:09 h2.limetransit.com vdsm-tool[3736]: exec_restore(cmd)
Dec 29 21:08:09 h2.limetransit.com vdsm-tool[3736]: File 
"/usr/lib/python2.7/site-packages/vdsm/tool/restore_nets.py", line 54, 
in exec_restore
Dec 29 21:08:09 h2.limetransit.com vdsm-tool[3736]: raise 
EnvironmentError('Failed to restore the persisted networks')
Dec 29 21:08:09 h2.limetransit.com vdsm-tool[3736]: EnvironmentError: 
Failed to restore the persisted networks

So I remove the nameservers section as well. At first I could not get it 
to work, as vdsm-network would find the backup file 
/var/lib/vdsm/persistence/netconf/nets/ovirtmgmt~ but after removing 
that it worked fine.

Not sure when nameservers and defaultroute was set in vdsm as it worked 
on reboot before.
/Sverker

Den 2016-12-29 kl. 14:20, skrev Sverker Abrahamsson:
>
> The specific command most likely fails because there is no chain named 
> libvirt-J-vnet0, but when should that have been created?
> /Sverker
>
> -------- Vidarebefordrat meddelande --------
> Ämne: 	Re: [ovirt-users] Issue with OVN/OVS and mandatory ovirtmgmt 
> network
> Datum: 	Thu, 29 Dec 2016 08:06:29 -0500 (EST)
> Från: 	Marcin Mirecki <mmirecki at redhat.com>
> Till: 	Sverker Abrahamsson <sverker at abrahamsson.com>
> Kopia: 	Ovirt Users <users at ovirt.org>, Lance Richardson 
> <lrichard at redhat.com>, Numan Siddique <nusiddiq at redhat.com>
>
>
>
> Let me add the OVN team.
>
> Lance, Numan,
>
> Can you please look at this?
>
> Trying to plug a vNIC results in:
> > >>>>>> Dec 28 23:31:35 h2 ovs-vsctl: ovs|00001|vsctl|INFO|Called as ovs-vsctl
> > >>>>>> --timeout=5 -- --if-exists del-port vnet0 -- add-port br-int vnet0 --
> > >>>>>> set Interface vnet0 "external-ids:attached-mac=\"00:1a:4a:16:01:51\""
> > >>>>>> -- set Interface vnet0
> > >>>>>> "external-ids:iface-id=\"e8853aac-8a75-41b0-8010-e630017dcdd8\"" --
> > >>>>>> set Interface vnet0
> > >>>>>> "external-ids:vm-id=\"b9440d60-ef5a-4e2b-83cf-081df7c09e6f\"" -- set
> > >>>>>> Interface vnet0 external-ids:iface-status=active
> > >>>>>> Dec 28 23:31:35 h2 kernel: device vnet0 entered promiscuous mode
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/ebtables --concurrent -t nat -D PREROUTING -i vnet0 -j
> > >>>>>> libvirt-J-vnet0' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
>
> More details below
>
>
> ----- Original Message -----
> > From: "Sverker Abrahamsson"<sverker at abrahamsson.com>
> > To: "Marcin Mirecki"<mmirecki at redhat.com>
> > Cc: "Ovirt Users"<users at ovirt.org>
> > Sent: Thursday, December 29, 2016 1:42:11 PM
> > Subject: Re: [ovirt-users] Issue with OVN/OVS and mandatory ovirtmgmt network
> >
> > Hi
> > Same problem still..
> > /Sverker
> >
> > Den 2016-12-29 kl. 13:34, skrev Marcin Mirecki:
> > > Hi,
> > >
> > > The tunnels are created to connect multiple OVN controllers.
> > > If there is only one, there is no need for the tunnels, so none
> > > will be created, this is the correct behavior.
> > >
> > > Does the problem still occur after setting configuring the OVN-controller?
> > >
> > > Marcin
> > >
> > > ----- Original Message -----
> > >> From: "Sverker Abrahamsson"<sverker at abrahamsson.com>
> > >> To: "Marcin Mirecki"<mmirecki at redhat.com>
> > >> Cc: "Ovirt Users"<users at ovirt.org>
> > >> Sent: Thursday, December 29, 2016 11:44:32 AM
> > >> Subject: Re: [ovirt-users] Issue with OVN/OVS and mandatory ovirtmgmt
> > >> network
> > >>
> > >> Hi
> > >> The rpm packages you listed in the other mail are installed but I had
> > >> not run vdsm-tool ovn-config to create tunnel as the OVN controller is
> > >> on the same host.
> > >>
> > >> [root at h2 ~]# rpm -q openvswitch-ovn-common
> > >> openvswitch-ovn-common-2.6.90-1.el7.centos.x86_64
> > >> [root at h2 ~]# rpm -q openvswitch-ovn-host
> > >> openvswitch-ovn-host-2.6.90-1.el7.centos.x86_64
> > >> [root at h2 ~]# rpm -q python-openvswitch
> > >> python-openvswitch-2.6.90-1.el7.centos.noarch
> > >>
> > >> After removing my manually created br-int and run
> > >>
> > >> vdsm-tool ovn-config 127.0.0.1 172.27.1.1
> > >>
> > >> then I have the br-int but 'ip link show' does not show any link
> > >> 'genev_sys_' nor does 'ovs-vsctl show' any port for ovn. I assume these
> > >> are when there is an actual tunnel?
> > >>
> > >> [root at h2 ~]# ovs-vsctl show
> > >> ebb6aede-cbbc-4f4f-a88a-a9cd72b2bd23
> > >>       Bridge br-int
> > >>           fail_mode: secure
> > >>           Port br-int
> > >>               Interface br-int
> > >>                   type: internal
> > >>       Bridge ovirtbridge
> > >>           Port ovirtbridge
> > >>               Interface ovirtbridge
> > >>                   type: internal
> > >>       Bridge "ovsbridge0"
> > >>           Port "ovsbridge0"
> > >>               Interface "ovsbridge0"
> > >>                   type: internal
> > >>           Port "eth0"
> > >>               Interface "eth0"
> > >>       ovs_version: "2.6.90"
> > >>
> > >> [root at h2 ~]# ip link show
> > >> 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN mode
> > >> DEFAULT qlen 1
> > >>       link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
> > >> 2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast
> > >> master ovs-system state UP mode DEFAULT qlen 1000
> > >>       link/ether 44:8a:5b:84:7d:b3 brd ff:ff:ff:ff:ff:ff
> > >> 3: ovs-system: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN mode
> > >> DEFAULT qlen 1000
> > >>       link/ether 5a:14:cf:28:47:e2 brd ff:ff:ff:ff:ff:ff
> > >> 4: ovsbridge0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue
> > >> state UNKNOWN mode DEFAULT qlen 1000
> > >>       link/ether 44:8a:5b:84:7d:b3 brd ff:ff:ff:ff:ff:ff
> > >> 5: br-int: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN mode
> > >> DEFAULT qlen 1000
> > >>       link/ether 9e:b0:3a:9d:f2:4b brd ff:ff:ff:ff:ff:ff
> > >> 6: ovirtbridge: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue
> > >> state UNKNOWN mode DEFAULT qlen 1000
> > >>       link/ether a6:f6:e5:a4:5b:45 brd ff:ff:ff:ff:ff:ff
> > >> 7: dummy0: <BROADCAST,NOARP,UP,LOWER_UP> mtu 1500 qdisc noqueue master
> > >> ovirtmgmt state UNKNOWN mode DEFAULT qlen 1000
> > >>       link/ether 66:e0:1c:c3:a9:d8 brd ff:ff:ff:ff:ff:ff
> > >> 8: ovirtmgmt: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue
> > >> state UP mode DEFAULT qlen 1000
> > >>       link/ether 66:e0:1c:c3:a9:d8 brd ff:ff:ff:ff:ff:ff
> > >>
> > >> Firewall settings:
> > >> [root at h2 ~]# firewall-cmd --list-all-zones
> > >> work
> > >>     target: default
> > >>     icmp-block-inversion: no
> > >>     interfaces:
> > >>     sources:
> > >>     services: dhcpv6-client ssh
> > >>     ports:
> > >>     protocols:
> > >>     masquerade: no
> > >>     forward-ports:
> > >>     sourceports:
> > >>     icmp-blocks:
> > >>     rich rules:
> > >>
> > >>
> > >> drop
> > >>     target: DROP
> > >>     icmp-block-inversion: no
> > >>     interfaces:
> > >>     sources:
> > >>     services:
> > >>     ports:
> > >>     protocols:
> > >>     masquerade: no
> > >>     forward-ports:
> > >>     sourceports:
> > >>     icmp-blocks:
> > >>     rich rules:
> > >>
> > >>
> > >> internal
> > >>     target: default
> > >>     icmp-block-inversion: no
> > >>     interfaces:
> > >>     sources:
> > >>     services: dhcpv6-client mdns samba-client ssh
> > >>     ports:
> > >>     protocols:
> > >>     masquerade: no
> > >>     forward-ports:
> > >>     sourceports:
> > >>     icmp-blocks:
> > >>     rich rules:
> > >>
> > >>
> > >> external
> > >>     target: default
> > >>     icmp-block-inversion: no
> > >>     interfaces:
> > >>     sources:
> > >>     services: ssh
> > >>     ports:
> > >>     protocols:
> > >>     masquerade: yes
> > >>     forward-ports:
> > >>     sourceports:
> > >>     icmp-blocks:
> > >>     rich rules:
> > >>
> > >>
> > >> trusted
> > >>     target: ACCEPT
> > >>     icmp-block-inversion: no
> > >>     interfaces:
> > >>     sources:
> > >>     services:
> > >>     ports:
> > >>     protocols:
> > >>     masquerade: no
> > >>     forward-ports:
> > >>     sourceports:
> > >>     icmp-blocks:
> > >>     rich rules:
> > >>
> > >>
> > >> home
> > >>     target: default
> > >>     icmp-block-inversion: no
> > >>     interfaces:
> > >>     sources:
> > >>     services: dhcpv6-client mdns samba-client ssh
> > >>     ports:
> > >>     protocols:
> > >>     masquerade: no
> > >>     forward-ports:
> > >>     sourceports:
> > >>     icmp-blocks:
> > >>     rich rules:
> > >>
> > >>
> > >> dmz
> > >>     target: default
> > >>     icmp-block-inversion: no
> > >>     interfaces:
> > >>     sources:
> > >>     services: ssh
> > >>     ports:
> > >>     protocols:
> > >>     masquerade: no
> > >>     forward-ports:
> > >>     sourceports:
> > >>     icmp-blocks:
> > >>     rich rules:
> > >>
> > >>
> > >> public (active)
> > >>     target: default
> > >>     icmp-block-inversion: no
> > >>     interfaces: eth0 ovsbridge0
> > >>     sources:
> > >>     services: dhcpv6-client ssh
> > >>     ports:
> > >>     protocols:
> > >>     masquerade: no
> > >>     forward-ports:
> > >>     sourceports:
> > >>     icmp-blocks:
> > >>     rich rules:
> > >>
> > >>
> > >> block
> > >>     target: %%REJECT%%
> > >>     icmp-block-inversion: no
> > >>     interfaces:
> > >>     sources:
> > >>     services:
> > >>     ports:
> > >>     protocols:
> > >>     masquerade: no
> > >>     forward-ports:
> > >>     sourceports:
> > >>     icmp-blocks:
> > >>     rich rules:
> > >>
> > >>
> > >> ovirt (active)
> > >>     target: default
> > >>     icmp-block-inversion: no
> > >>     interfaces: ovirtbridge ovirtmgmt
> > >>     sources:
> > >>     services: dhcp ovirt-fence-kdump-listener ovirt-http ovirt-https
> > >> ovirt-imageio-proxy ovirt-postgres ovirt-provider-ovn
> > >> ovirt-vmconsole-proxy ovirt-websocket-proxy ssh vdsm
> > >>     ports:
> > >>     protocols:
> > >>     masquerade: yes
> > >>     forward-ports:
> > >>     sourceports:
> > >>     icmp-blocks:
> > >>     rich rules:
> > >>           rule family="ipv4" port port="6641" protocol="tcp" accept
> > >>           rule family="ipv4" port port="6642" protocol="tcp" accept
> > >>
> > >> The db dump is attached
> > >> /Sverker
> > >> Den 2016-12-29 kl. 09:50, skrev Marcin Mirecki:
> > >>> Hi,
> > >>>
> > >>> Can you please do: "sudo ovsdb-client dump"
> > >>> on the host and send me the output?
> > >>>
> > >>> Have you configured the ovn controller to connect to the
> > >>> OVN north? You can do it using "vdsm-tool ovn-config" or
> > >>> using the OVN tools directly.
> > >>> Please check out:https://www.ovirt.org/blog/2016/11/ovirt-provider-ovn/
> > >>> for details.
> > >>>
> > >>> Also please note that the OVN provider is completely different
> > >>> from the neutron-openvswitch plugin. Please don't mix the two.
> > >>>
> > >>> Marcin
> > >>>
> > >>>
> > >>> ----- Original Message -----
> > >>>> From: "Marcin Mirecki"<mmirecki at redhat.com>
> > >>>> To: "Sverker Abrahamsson"<sverker at abrahamsson.com>
> > >>>> Cc: "Ovirt Users"<users at ovirt.org>
> > >>>> Sent: Thursday, December 29, 2016 9:27:19 AM
> > >>>> Subject: Re: [ovirt-users] Issue with OVN/OVS and mandatory ovirtmgmt
> > >>>> network
> > >>>>
> > >>>> Hi,
> > >>>>
> > >>>> br-int is the OVN integration bridge, it should have been created
> > >>>> when installing OVN. I assume you have the following packages installed
> > >>>> on the host:
> > >>>>       openvswitch-ovn-common
> > >>>>       openvswitch-ovn-host
> > >>>>       python-openvswitch
> > >>>>
> > >>>> Please give me some time to look at the connectivity problem.
> > >>>>
> > >>>> Marcin
> > >>>>
> > >>>>
> > >>>>
> > >>>> ----- Original Message -----
> > >>>>> From: "Sverker Abrahamsson"<sverker at abrahamsson.com>
> > >>>>> To: "Marcin Mirecki"<mmirecki at redhat.com>
> > >>>>> Cc: "Ovirt Users"<users at ovirt.org>
> > >>>>> Sent: Thursday, December 29, 2016 12:47:04 AM
> > >>>>> Subject: Re: [ovirt-users] Issue with OVN/OVS and mandatory ovirtmgmt
> > >>>>> network
> > >>>>>
> > >>>>> From
> > >>>>> /usr/libexec/vdsm/hooks/before_device_create/ovirt_provider_ovn_hook
> > >>>>> (installed by ovirt-provider-ovn-driver rpm):
> > >>>>>
> > >>>>> BRIDGE_NAME = 'br-int'
> > >>>>>
> > >>>>>
> > >>>>> Den 2016-12-28 kl. 23:56, skrev Sverker Abrahamsson:
> > >>>>>> Googling on the message about br-int suggested adding that bridge to
> > >>>>>> ovs:
> > >>>>>>
> > >>>>>> ovs-vsctl add-br br-int
> > >>>>>>
> > >>>>>> Then the VM is able to boot, but it fails to get network connectivity.
> > >>>>>> Output in /var/log/messages:
> > >>>>>>
> > >>>>>> Dec 28 23:31:35 h2 ovs-vsctl: ovs|00001|vsctl|INFO|Called as ovs-vsctl
> > >>>>>> --timeout=5 -- --if-exists del-port vnet0 -- add-port br-int vnet0 --
> > >>>>>> set Interface vnet0 "external-ids:attached-mac=\"00:1a:4a:16:01:51\""
> > >>>>>> -- set Interface vnet0
> > >>>>>> "external-ids:iface-id=\"e8853aac-8a75-41b0-8010-e630017dcdd8\"" --
> > >>>>>> set Interface vnet0
> > >>>>>> "external-ids:vm-id=\"b9440d60-ef5a-4e2b-83cf-081df7c09e6f\"" -- set
> > >>>>>> Interface vnet0 external-ids:iface-status=active
> > >>>>>> Dec 28 23:31:35 h2 kernel: device vnet0 entered promiscuous mode
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/ebtables --concurrent -t nat -D PREROUTING -i vnet0 -j
> > >>>>>> libvirt-J-vnet0' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/ebtables --concurrent -t nat -D POSTROUTING -o vnet0 -j
> > >>>>>> libvirt-P-vnet0' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/ebtables --concurrent -t nat -L libvirt-J-vnet0' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/ebtables --concurrent -t nat -L libvirt-P-vnet0' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/ebtables --concurrent -t nat -F libvirt-J-vnet0' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/ebtables --concurrent -t nat -X libvirt-J-vnet0' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/ebtables --concurrent -t nat -F libvirt-P-vnet0' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/ebtables --concurrent -t nat -X libvirt-P-vnet0' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/ebtables --concurrent -t nat -F J-vnet0-mac' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/ebtables --concurrent -t nat -X J-vnet0-mac' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/ebtables --concurrent -t nat -F J-vnet0-arp-mac' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/ebtables --concurrent -t nat -X J-vnet0-arp-mac' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/iptables -w2 -w -D libvirt-out -m physdev
> > >>>>>> --physdev-is-bridged --physdev-out vnet0 -g FO-vnet0' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/iptables -w2 -w -D libvirt-out -m physdev --physdev-out
> > >>>>>> vnet0 -g FO-vnet0' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/iptables -w2 -w -D libvirt-in -m physdev --physdev-in vnet0
> > >>>>>> -g FI-vnet0' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/iptables -w2 -w -D libvirt-host-in -m physdev --physdev-in
> > >>>>>> vnet0 -g HI-vnet0' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/iptables -w2 -w -F FO-vnet0' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/iptables -w2 -w -X FO-vnet0' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/iptables -w2 -w -F FI-vnet0' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/iptables -w2 -w -X FI-vnet0' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/iptables -w2 -w -F HI-vnet0' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/iptables -w2 -w -X HI-vnet0' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/iptables -w2 -w -E FP-vnet0 FO-vnet0' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/iptables -w2 -w -E FJ-vnet0 FI-vnet0' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/iptables -w2 -w -E HJ-vnet0 HI-vnet0' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/ip6tables -w2 -w -D libvirt-out -m physdev
> > >>>>>> --physdev-is-bridged --physdev-out vnet0 -g FO-vnet0' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/ip6tables -w2 -w -D libvirt-out -m physdev --physdev-out
> > >>>>>> vnet0 -g FO-vnet0' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/ip6tables -w2 -w -D libvirt-in -m physdev --physdev-in
> > >>>>>> vnet0 -g FI-vnet0' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/ip6tables -w2 -w -D libvirt-host-in -m physdev --physdev-in
> > >>>>>> vnet0 -g HI-vnet0' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/ip6tables -w2 -w -F FO-vnet0' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/ip6tables -w2 -w -X FO-vnet0' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/ip6tables -w2 -w -F FI-vnet0' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/ip6tables -w2 -w -X FI-vnet0' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/ip6tables -w2 -w -F HI-vnet0' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/ip6tables -w2 -w -X HI-vnet0' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/ip6tables -w2 -w -E FP-vnet0 FO-vnet0' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/ip6tables -w2 -w -E FJ-vnet0 FI-vnet0' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/ip6tables -w2 -w -E HJ-vnet0 HI-vnet0' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/ebtables --concurrent -t nat -D PREROUTING -i vnet0 -j
> > >>>>>> libvirt-I-vnet0' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/ebtables --concurrent -t nat -D POSTROUTING -o vnet0 -j
> > >>>>>> libvirt-O-vnet0' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/ebtables --concurrent -t nat -L libvirt-I-vnet0' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/ebtables --concurrent -t nat -L libvirt-O-vnet0' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/ebtables --concurrent -t nat -F libvirt-I-vnet0' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/ebtables --concurrent -t nat -X libvirt-I-vnet0' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/ebtables --concurrent -t nat -F libvirt-O-vnet0' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/ebtables --concurrent -t nat -X libvirt-O-vnet0' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/ebtables --concurrent -t nat -L libvirt-P-vnet0' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/ebtables --concurrent -t nat -E libvirt-P-vnet0
> > >>>>>> libvirt-O-vnet0' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/ebtables --concurrent -t nat -F I-vnet0-mac' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/ebtables --concurrent -t nat -X I-vnet0-mac' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/ebtables --concurrent -t nat -F I-vnet0-arp-mac' failed:
> > >>>>>> Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED:
> > >>>>>> '/usr/sbin/ebtables --concurrent -t nat -X I-vnet0-arp-mac' failed:
> > >>>>>>
> > >>>>>>
> > >>>>>> [root at h2 etc]# ovs-vsctl show
> > >>>>>> ebb6aede-cbbc-4f4f-a88a-a9cd72b2bd23
> > >>>>>>       Bridge ovirtbridge
> > >>>>>>           Port "ovirtport0"
> > >>>>>>               Interface "ovirtport0"
> > >>>>>>                   type: internal
> > >>>>>>           Port ovirtbridge
> > >>>>>>               Interface ovirtbridge
> > >>>>>>                   type: internal
> > >>>>>>       Bridge "ovsbridge0"
> > >>>>>>           Port "ovsbridge0"
> > >>>>>>               Interface "ovsbridge0"
> > >>>>>>                   type: internal
> > >>>>>>           Port "eth0"
> > >>>>>>               Interface "eth0"
> > >>>>>>       Bridge br-int
> > >>>>>>           Port br-int
> > >>>>>>               Interface br-int
> > >>>>>>                   type: internal
> > >>>>>>           Port "vnet0"
> > >>>>>>               Interface "vnet0"
> > >>>>>>       ovs_version: "2.6.90"
> > >>>>>>
> > >>>>>> Searching through the code it appears that br-int comes from
> > >>>>>> neutron-openvswitch plugin ??
> > >>>>>>
> > >>>>>> [root at h2 share]# rpm -qf
> > >>>>>> /usr/share/otopi/plugins/ovirt-host-deploy/openstack/neutron_openvswitch.py
> > >>>>>> ovirt-host-deploy-1.6.0-0.0.master.20161215101008.gitb76ad50.el7.centos.noarch
> > >>>>>>
> > >>>>>>
> > >>>>>> /Sverker
> > >>>>>>
> > >>>>>> Den 2016-12-28 kl. 23:24, skrev Sverker Abrahamsson:
> > >>>>>>> In addition I had to add an alias to modprobe:
> > >>>>>>>
> > >>>>>>> [root at h2 modprobe.d]# cat dummy.conf
> > >>>>>>> alias dummy0 dummy
> > >>>>>>>
> > >>>>>>>
> > >>>>>>> Den 2016-12-28 kl. 23:03, skrev Sverker Abrahamsson:
> > >>>>>>>> Hi
> > >>>>>>>> I first tried to set device name to dummy_0, but then ifup did not
> > >>>>>>>> succeed in creating the device unless I first did 'ip link add
> > >>>>>>>> dummy_0 type dummy' but then it would not suceed to establish the if
> > >>>>>>>> on reboot.
> > >>>>>>>>
> > >>>>>>>> Setting fake_nics = dummy0 would not work neither, but this works:
> > >>>>>>>>
> > >>>>>>>> fake_nics = dummy*
> > >>>>>>>>
> > >>>>>>>> The engine is now able to find the if and assign bridge ovirtmgmt to
> > >>>>>>>> it.
> > >>>>>>>>
> > >>>>>>>> However, I then run into the next issue when starting a VM:
> > >>>>>>>>
> > >>>>>>>> 2016-12-28 22:28:23,897 ERROR
> > >>>>>>>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
> > >>>>>>>> (ForkJoinPool-1-worker-2) [] Correlation ID: null, Call Stack: null,
> > >>>>>>>> Custom Event ID: -1, Message: VM CentOS7 is down with error. Exit
> > >>>>>>>> message: Cannot get interface MTU on 'br-int': No such device.
> > >>>>>>>>
> > >>>>>>>> This VM has a nic on ovirtbridge, which comes from the OVN provider.
> > >>>>>>>>
> > >>>>>>>> /Sverker
> > >>>>>>>>
> > >>>>>>>> Den 2016-12-28 kl. 14:38, skrev Marcin Mirecki:
> > >>>>>>>>> Sverker,
> > >>>>>>>>>
> > >>>>>>>>> Can you try adding a vnic named veth_* or dummy_*,
> > >>>>>>>>> (or alternatively add the name of the vnic to
> > >>>>>>>>> vdsm.config fake_nics), and setup the management
> > >>>>>>>>> network using this vnic?
> > >>>>>>>>> I suppose adding the vnic you use for connecting
> > >>>>>>>>> to the engine to fake_nics should make it visible
> > >>>>>>>>> to the engine, and you should be able to use it for
> > >>>>>>>>> the setup.
> > >>>>>>>>>
> > >>>>>>>>> Marcin
> > >>>>>>>>>
> > >>>>>>>>>
> > >>>>>>>>>
> > >>>>>>>>> ----- Original Message -----
> > >>>>>>>>>> From: "Marcin Mirecki"<mmirecki at redhat.com>
> > >>>>>>>>>> To: "Sverker Abrahamsson"<sverker at abrahamsson.com>
> > >>>>>>>>>> Cc: "Ovirt Users"<users at ovirt.org>
> > >>>>>>>>>> Sent: Wednesday, December 28, 2016 12:06:26 PM
> > >>>>>>>>>> Subject: Re: [ovirt-users] Issue with OVN/OVS and mandatory
> > >>>>>>>>>> ovirtmgmt network
> > >>>>>>>>>>
> > >>>>>>>>>>> I have an internal OVS bridge called ovirtbridge which has a port
> > >>>>>>>>>>> with
> > >>>>>>>>>>> IP address, but in the host network settings that port is not
> > >>>>>>>>>>> visible.
> > >>>>>>>>>> I just verified and unfortunately the virtual ports are not
> > >>>>>>>>>> visible in engine
> > >>>>>>>>>> to assign a network to :(
> > >>>>>>>>>> I'm afraid that the engine is not ready for such a scenario (even
> > >>>>>>>>>> if it
> > >>>>>>>>>> works).
> > >>>>>>>>>> Please give me some time to look for a solution.
> > >>>>>>>>>>
> > >>>>>>>>>> ----- Original Message -----
> > >>>>>>>>>>> From: "Sverker Abrahamsson"<sverker at abrahamsson.com>
> > >>>>>>>>>>> To: "Marcin Mirecki"<mmirecki at redhat.com>
> > >>>>>>>>>>> Cc: "Ovirt Users"<users at ovirt.org>
> > >>>>>>>>>>> Sent: Wednesday, December 28, 2016 11:48:24 AM
> > >>>>>>>>>>> Subject: Re: [ovirt-users] Issue with OVN/OVS and mandatory
> > >>>>>>>>>>> ovirtmgmt
> > >>>>>>>>>>> network
> > >>>>>>>>>>>
> > >>>>>>>>>>> Hi Marcin
> > >>>>>>>>>>> Yes, that is my issue. I don't want to let ovirt/vdsm see eth0
> > >>>>>>>>>>> nor
> > >>>>>>>>>>> ovsbridge0 since as soon as it sees them it messes up the network
> > >>>>>>>>>>> config
> > >>>>>>>>>>> so that the host will be unreachable.
> > >>>>>>>>>>>
> > >>>>>>>>>>> I have an internal OVS bridge called ovirtbridge which has a port
> > >>>>>>>>>>> with
> > >>>>>>>>>>> IP address, but in the host network settings that port is not
> > >>>>>>>>>>> visible.
> > >>>>>>>>>>> It doesn't help to name it ovirtmgmt.
> > >>>>>>>>>>>
> > >>>>>>>>>>> The engine is able to communicate with the host on the ip it has
> > >>>>>>>>>>> been
> > >>>>>>>>>>> given, it's just that it believes that it HAS to have a ovirtmgmt
> > >>>>>>>>>>> network which can't be on OVN.
> > >>>>>>>>>>>
> > >>>>>>>>>>> /Sverker
> > >>>>>>>>>>>
> > >>>>>>>>>>>
> > >>>>>>>>>>> Den 2016-12-28 kl. 10:45, skrev Marcin Mirecki:
> > >>>>>>>>>>>> Hi Sverker,
> > >>>>>>>>>>>>
> > >>>>>>>>>>>> The management network is mandatory on each host. It's used by
> > >>>>>>>>>>>> the
> > >>>>>>>>>>>> engine to communicate with the host.
> > >>>>>>>>>>>> Looking at your description and the exception it looks like it
> > >>>>>>>>>>>> is
> > >>>>>>>>>>>> missing.
> > >>>>>>>>>>>> The error is caused by not having any network for the host
> > >>>>>>>>>>>> (network list retrieved in
> > >>>>>>>>>>>> InterfaceDaoImpl.getHostNetworksByCluster -
> > >>>>>>>>>>>> which
> > >>>>>>>>>>>> gets all the networks on nics for a host from vds_interface
> > >>>>>>>>>>>> table in the
> > >>>>>>>>>>>> DB).
> > >>>>>>>>>>>>
> > >>>>>>>>>>>> Could you maybe create a virtual nic connected to ovsbridge0 (as
> > >>>>>>>>>>>> I
> > >>>>>>>>>>>> understand you
> > >>>>>>>>>>>> have no physical nic available) and use this for the management
> > >>>>>>>>>>>> network?
> > >>>>>>>>>>>>
> > >>>>>>>>>>>>> I then create a bridge for use with ovirt, with a private
> > >>>>>>>>>>>>> address.
> > >>>>>>>>>>>> I'm not quite sure I understand. Is this yet another bridge
> > >>>>>>>>>>>> connected to
> > >>>>>>>>>>>> ovsbridge0?
> > >>>>>>>>>>>> You could also attach the vnic for the management network here
> > >>>>>>>>>>>> if need
> > >>>>>>>>>>>> be.
> > >>>>>>>>>>>>
> > >>>>>>>>>>>> Please keep in mind that OVN has no use in setting up the
> > >>>>>>>>>>>> management
> > >>>>>>>>>>>> network.
> > >>>>>>>>>>>> The OVN provider can only handle external networks, which can
> > >>>>>>>>>>>> not be used
> > >>>>>>>>>>>> for a
> > >>>>>>>>>>>> management network.
> > >>>>>>>>>>>>
> > >>>>>>>>>>>> Marcin
> > >>>>>>>>>>>>
> > >>>>>>>>>>>>
> > >>>>>>>>>>>> ----- Original Message -----
> > >>>>>>>>>>>>> From: "Sverker Abrahamsson"<sverker at abrahamsson.com>
> > >>>>>>>>>>>>> To:users at ovirt.org
> > >>>>>>>>>>>>> Sent: Wednesday, December 28, 2016 12:39:59 AM
> > >>>>>>>>>>>>> Subject: [ovirt-users] Issue with OVN/OVS and mandatory
> > >>>>>>>>>>>>> ovirtmgmt
> > >>>>>>>>>>>>> network
> > >>>>>>>>>>>>>
> > >>>>>>>>>>>>>
> > >>>>>>>>>>>>>
> > >>>>>>>>>>>>> Hi
> > >>>>>>>>>>>>> For long time I've been looking for proper support in ovirt for
> > >>>>>>>>>>>>> Open
> > >>>>>>>>>>>>> vSwitch
> > >>>>>>>>>>>>> so I'm happy that it is moving in the right direction. However,
> > >>>>>>>>>>>>> there
> > >>>>>>>>>>>>> seems
> > >>>>>>>>>>>>> to still be a dependency on a ovirtmgmt bridge and I'm unable
> > >>>>>>>>>>>>> to move
> > >>>>>>>>>>>>> that
> > >>>>>>>>>>>>> to the OVN provider.
> > >>>>>>>>>>>>>
> > >>>>>>>>>>>>> The hosting center where I rent hw instances has a bit special
> > >>>>>>>>>>>>> network
> > >>>>>>>>>>>>> setup,
> > >>>>>>>>>>>>> so I have one physical network port with a /32 netmask and
> > >>>>>>>>>>>>> point-to-point
> > >>>>>>>>>>>>> config to router. The physical port I connect to a ovs bridge
> > >>>>>>>>>>>>> which has
> > >>>>>>>>>>>>> the
> > >>>>>>>>>>>>> public ip. Since ovirt always messes up the network config when
> > >>>>>>>>>>>>> I've
> > >>>>>>>>>>>>> tried
> > >>>>>>>>>>>>> to let it have access to the network config for the physical
> > >>>>>>>>>>>>> port, I've
> > >>>>>>>>>>>>> set
> > >>>>>>>>>>>>> eht0 and ovsbridge0 as hidden in vdsm.conf.
> > >>>>>>>>>>>>>
> > >>>>>>>>>>>>>
> > >>>>>>>>>>>>> I then create a bridge for use with ovirt, with a private
> > >>>>>>>>>>>>> address. With
> > >>>>>>>>>>>>> the
> > >>>>>>>>>>>>> OVN provider I am now able to import these into the engine and
> > >>>>>>>>>>>>> it looks
> > >>>>>>>>>>>>> good. When creating a VM I can select that it will have a vNic
> > >>>>>>>>>>>>> on my OVS
> > >>>>>>>>>>>>> bridge.
> > >>>>>>>>>>>>>
> > >>>>>>>>>>>>> However, I can't start the VM as an exception is thrown in the
> > >>>>>>>>>>>>> log:
> > >>>>>>>>>>>>>
> > >>>>>>>>>>>>> 2016-12-28 00:13:33,350 ERROR
> > >>>>>>>>>>>>> [org.ovirt.engine.core.bll.RunVmCommand]
> > >>>>>>>>>>>>> (default task-5) [3c882d53] Error during ValidateFailure.:
> > >>>>>>>>>>>>> java.lang.NullPointerException
> > >>>>>>>>>>>>> at
> > >>>>>>>>>>>>> org.ovirt.engine.core.bll.scheduling.policyunits.NetworkPolicyUnit.validateRequiredNetworksAvailable(NetworkPolicyUnit.java:140)
> > >>>>>>>>>>>>>
> > >>>>>>>>>>>>> [bll.jar:]
> > >>>>>>>>>>>>> at
> > >>>>>>>>>>>>> org.ovirt.engine.core.bll.scheduling.policyunits.NetworkPolicyUnit.filter(NetworkPolicyUnit.java:69)
> > >>>>>>>>>>>>>
> > >>>>>>>>>>>>> [bll.jar:]
> > >>>>>>>>>>>>> at
> > >>>>>>>>>>>>> org.ovirt.engine.core.bll.scheduling.SchedulingManager.runInternalFilters(SchedulingManager.java:597)
> > >>>>>>>>>>>>>
> > >>>>>>>>>>>>> [bll.jar:]
> > >>>>>>>>>>>>> at
> > >>>>>>>>>>>>> org.ovirt.engine.core.bll.scheduling.SchedulingManager.runFilters(SchedulingManager.java:564)
> > >>>>>>>>>>>>>
> > >>>>>>>>>>>>> [bll.jar:]
> > >>>>>>>>>>>>> at
> > >>>>>>>>>>>>> org.ovirt.engine.core.bll.scheduling.SchedulingManager.canSchedule(SchedulingManager.java:494)
> > >>>>>>>>>>>>>
> > >>>>>>>>>>>>> [bll.jar:]
> > >>>>>>>>>>>>> at
> > >>>>>>>>>>>>> org.ovirt.engine.core.bll.validator.RunVmValidator.canRunVm(RunVmValidator.java:133)
> > >>>>>>>>>>>>>
> > >>>>>>>>>>>>> [bll.jar:]
> > >>>>>>>>>>>>> at
> > >>>>>>>>>>>>> org.ovirt.engine.core.bll.RunVmCommand.validate(RunVmCommand.java:940)
> > >>>>>>>>>>>>>
> > >>>>>>>>>>>>> [bll.jar:]
> > >>>>>>>>>>>>> at
> > >>>>>>>>>>>>> org.ovirt.engine.core.bll.CommandBase.internalValidate(CommandBase.java:886)
> > >>>>>>>>>>>>>
> > >>>>>>>>>>>>> [bll.jar:]
> > >>>>>>>>>>>>> at
> > >>>>>>>>>>>>> org.ovirt.engine.core.bll.CommandBase.validateOnly(CommandBase.java:366)
> > >>>>>>>>>>>>>
> > >>>>>>>>>>>>> [bll.jar:]
> > >>>>>>>>>>>>> at
> > >>>>>>>>>>>>> org.ovirt.engine.core.bll.PrevalidatingMultipleActionsRunner.canRunActions(PrevalidatingMultipleActionsRunner.java:113)
> > >>>>>>>>>>>>>
> > >>>>>>>>>>>>> [bll.jar:]
> > >>>>>>>>>>>>> at
> > >>>>>>>>>>>>> org.ovirt.engine.core.bll.PrevalidatingMultipleActionsRunner.invokeCommands(PrevalidatingMultipleActionsRunner.java:99)
> > >>>>>>>>>>>>>
> > >>>>>>>>>>>>> [bll.jar:]
> > >>>>>>>>>>>>> at
> > >>>>>>>>>>>>> org.ovirt.engine.core.bll.PrevalidatingMultipleActionsRunner.execute(PrevalidatingMultipleActionsRunner.java:76)
> > >>>>>>>>>>>>>
> > >>>>>>>>>>>>> [bll.jar:]
> > >>>>>>>>>>>>> at
> > >>>>>>>>>>>>> org.ovirt.engine.core.bll.Backend.runMultipleActionsImpl(Backend.java:613)
> > >>>>>>>>>>>>>
> > >>>>>>>>>>>>> [bll.jar:]
> > >>>>>>>>>>>>> at
> > >>>>>>>>>>>>> org.ovirt.engine.core.bll.Backend.runMultipleActions(Backend.java:583)
> > >>>>>>>>>>>>>
> > >>>>>>>>>>>>> [bll.jar:]
> > >>>>>>>>>>>>>
> > >>>>>>>>>>>>>
> > >>>>>>>>>>>>> Looking at that section of code where the exception is thrown,
> > >>>>>>>>>>>>> I see
> > >>>>>>>>>>>>> that
> > >>>>>>>>>>>>> it
> > >>>>>>>>>>>>> iterates over host networks to find required networks, which I
> > >>>>>>>>>>>>> assume is
> > >>>>>>>>>>>>> ovirtmgmt. In the host network setup dialog I don't see any
> > >>>>>>>>>>>>> networks at
> > >>>>>>>>>>>>> all
> > >>>>>>>>>>>>> but it lists ovirtmgmt as required. It also list the OVN
> > >>>>>>>>>>>>> networks but
> > >>>>>>>>>>>>> these
> > >>>>>>>>>>>>> can't be statically assigned as they are added dynamically when
> > >>>>>>>>>>>>> needed,
> > >>>>>>>>>>>>> which is fine.
> > >>>>>>>>>>>>>
> > >>>>>>>>>>>>> I believe that I either need to remove ovirtmgmt network or
> > >>>>>>>>>>>>> configure
> > >>>>>>>>>>>>> that
> > >>>>>>>>>>>>> it
> > >>>>>>>>>>>>> is provided by the OVN provider, but neither is possible.
> > >>>>>>>>>>>>> Preferably it
> > >>>>>>>>>>>>> shouldn't be hardcoded which network is management and
> > >>>>>>>>>>>>> mandatory but be
> > >>>>>>>>>>>>> possible to configure.
> > >>>>>>>>>>>>>
> > >>>>>>>>>>>>> /Sverker
> > >>>>>>>>>>>>> Den 2016-12-27 kl. 17:10, skrev Marcin Mirecki:
> > >>>>>>>>>>>>>
> > >>>>>>>>>>>>>
> > >>>>>>>>>> _______________________________________________
> > >>>>>>>>>> Users mailing list
> > >>>>>>>>>>Users at ovirt.org
> > >>>>>>>>>>http://lists.ovirt.org/mailman/listinfo/users
> > >>>>>>>>>>
> > >>>>>>>> _______________________________________________
> > >>>>>>>> Users mailing list
> > >>>>>>>>Users at ovirt.org
> > >>>>>>>>http://lists.ovirt.org/mailman/listinfo/users
> > >>>>>>> _______________________________________________
> > >>>>>>> Users mailing list
> > >>>>>>>Users at ovirt.org
> > >>>>>>>http://lists.ovirt.org/mailman/listinfo/users
> > >>>>>> _______________________________________________
> > >>>>>> Users mailing list
> > >>>>>>Users at ovirt.org
> > >>>>>>http://lists.ovirt.org/mailman/listinfo/users
> > >>>> _______________________________________________
> > >>>> Users mailing list
> > >>>>Users at ovirt.org
> > >>>>http://lists.ovirt.org/mailman/listinfo/users
> > >>>>
> > >>
> >
> >
>
>
> _______________________________________________
> Users mailing list
> Users at ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20161229/d0705728/attachment-0001.html>


More information about the Users mailing list