
Hi The rpm packages you listed in the other mail are installed but I had not run vdsm-tool ovn-config to create tunnel as the OVN controller is on the same host. [root@h2 ~]# rpm -q openvswitch-ovn-common openvswitch-ovn-common-2.6.90-1.el7.centos.x86_64 [root@h2 ~]# rpm -q openvswitch-ovn-host openvswitch-ovn-host-2.6.90-1.el7.centos.x86_64 [root@h2 ~]# rpm -q python-openvswitch python-openvswitch-2.6.90-1.el7.centos.noarch After removing my manually created br-int and run vdsm-tool ovn-config 127.0.0.1 172.27.1.1 then I have the br-int but 'ip link show' does not show any link 'genev_sys_' nor does 'ovs-vsctl show' any port for ovn. I assume these are when there is an actual tunnel? [root@h2 ~]# ovs-vsctl show ebb6aede-cbbc-4f4f-a88a-a9cd72b2bd23 Bridge br-int fail_mode: secure Port br-int Interface br-int type: internal Bridge ovirtbridge Port ovirtbridge Interface ovirtbridge type: internal Bridge "ovsbridge0" Port "ovsbridge0" Interface "ovsbridge0" type: internal Port "eth0" Interface "eth0" ovs_version: "2.6.90" [root@h2 ~]# ip link show 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN mode DEFAULT qlen 1 link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast master ovs-system state UP mode DEFAULT qlen 1000 link/ether 44:8a:5b:84:7d:b3 brd ff:ff:ff:ff:ff:ff 3: ovs-system: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN mode DEFAULT qlen 1000 link/ether 5a:14:cf:28:47:e2 brd ff:ff:ff:ff:ff:ff 4: ovsbridge0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state UNKNOWN mode DEFAULT qlen 1000 link/ether 44:8a:5b:84:7d:b3 brd ff:ff:ff:ff:ff:ff 5: br-int: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN mode DEFAULT qlen 1000 link/ether 9e:b0:3a:9d:f2:4b brd ff:ff:ff:ff:ff:ff 6: ovirtbridge: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state UNKNOWN mode DEFAULT qlen 1000 link/ether a6:f6:e5:a4:5b:45 brd ff:ff:ff:ff:ff:ff 7: dummy0: <BROADCAST,NOARP,UP,LOWER_UP> mtu 1500 qdisc noqueue master ovirtmgmt state UNKNOWN mode DEFAULT qlen 1000 link/ether 66:e0:1c:c3:a9:d8 brd ff:ff:ff:ff:ff:ff 8: ovirtmgmt: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state UP mode DEFAULT qlen 1000 link/ether 66:e0:1c:c3:a9:d8 brd ff:ff:ff:ff:ff:ff Firewall settings: [root@h2 ~]# firewall-cmd --list-all-zones work target: default icmp-block-inversion: no interfaces: sources: services: dhcpv6-client ssh ports: protocols: masquerade: no forward-ports: sourceports: icmp-blocks: rich rules: drop target: DROP icmp-block-inversion: no interfaces: sources: services: ports: protocols: masquerade: no forward-ports: sourceports: icmp-blocks: rich rules: internal target: default icmp-block-inversion: no interfaces: sources: services: dhcpv6-client mdns samba-client ssh ports: protocols: masquerade: no forward-ports: sourceports: icmp-blocks: rich rules: external target: default icmp-block-inversion: no interfaces: sources: services: ssh ports: protocols: masquerade: yes forward-ports: sourceports: icmp-blocks: rich rules: trusted target: ACCEPT icmp-block-inversion: no interfaces: sources: services: ports: protocols: masquerade: no forward-ports: sourceports: icmp-blocks: rich rules: home target: default icmp-block-inversion: no interfaces: sources: services: dhcpv6-client mdns samba-client ssh ports: protocols: masquerade: no forward-ports: sourceports: icmp-blocks: rich rules: dmz target: default icmp-block-inversion: no interfaces: sources: services: ssh ports: protocols: masquerade: no forward-ports: sourceports: icmp-blocks: rich rules: public (active) target: default icmp-block-inversion: no interfaces: eth0 ovsbridge0 sources: services: dhcpv6-client ssh ports: protocols: masquerade: no forward-ports: sourceports: icmp-blocks: rich rules: block target: %%REJECT%% icmp-block-inversion: no interfaces: sources: services: ports: protocols: masquerade: no forward-ports: sourceports: icmp-blocks: rich rules: ovirt (active) target: default icmp-block-inversion: no interfaces: ovirtbridge ovirtmgmt sources: services: dhcp ovirt-fence-kdump-listener ovirt-http ovirt-https ovirt-imageio-proxy ovirt-postgres ovirt-provider-ovn ovirt-vmconsole-proxy ovirt-websocket-proxy ssh vdsm ports: protocols: masquerade: yes forward-ports: sourceports: icmp-blocks: rich rules: rule family="ipv4" port port="6641" protocol="tcp" accept rule family="ipv4" port port="6642" protocol="tcp" accept The db dump is attached /Sverker Den 2016-12-29 kl. 09:50, skrev Marcin Mirecki:
Hi,
Can you please do: "sudo ovsdb-client dump" on the host and send me the output?
Have you configured the ovn controller to connect to the OVN north? You can do it using "vdsm-tool ovn-config" or using the OVN tools directly. Please check out: https://www.ovirt.org/blog/2016/11/ovirt-provider-ovn/ for details.
Also please note that the OVN provider is completely different from the neutron-openvswitch plugin. Please don't mix the two.
Marcin
----- Original Message -----
From: "Marcin Mirecki" <mmirecki@redhat.com> To: "Sverker Abrahamsson" <sverker@abrahamsson.com> Cc: "Ovirt Users" <users@ovirt.org> Sent: Thursday, December 29, 2016 9:27:19 AM Subject: Re: [ovirt-users] Issue with OVN/OVS and mandatory ovirtmgmt network
Hi,
br-int is the OVN integration bridge, it should have been created when installing OVN. I assume you have the following packages installed on the host: openvswitch-ovn-common openvswitch-ovn-host python-openvswitch
Please give me some time to look at the connectivity problem.
Marcin
----- Original Message -----
From: "Sverker Abrahamsson" <sverker@abrahamsson.com> To: "Marcin Mirecki" <mmirecki@redhat.com> Cc: "Ovirt Users" <users@ovirt.org> Sent: Thursday, December 29, 2016 12:47:04 AM Subject: Re: [ovirt-users] Issue with OVN/OVS and mandatory ovirtmgmt network
From /usr/libexec/vdsm/hooks/before_device_create/ovirt_provider_ovn_hook (installed by ovirt-provider-ovn-driver rpm):
BRIDGE_NAME = 'br-int'
Den 2016-12-28 kl. 23:56, skrev Sverker Abrahamsson:
Googling on the message about br-int suggested adding that bridge to ovs:
ovs-vsctl add-br br-int
Then the VM is able to boot, but it fails to get network connectivity. Output in /var/log/messages:
Dec 28 23:31:35 h2 ovs-vsctl: ovs|00001|vsctl|INFO|Called as ovs-vsctl --timeout=5 -- --if-exists del-port vnet0 -- add-port br-int vnet0 -- set Interface vnet0 "external-ids:attached-mac=\"00:1a:4a:16:01:51\"" -- set Interface vnet0 "external-ids:iface-id=\"e8853aac-8a75-41b0-8010-e630017dcdd8\"" -- set Interface vnet0 "external-ids:vm-id=\"b9440d60-ef5a-4e2b-83cf-081df7c09e6f\"" -- set Interface vnet0 external-ids:iface-status=active Dec 28 23:31:35 h2 kernel: device vnet0 entered promiscuous mode Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/ebtables --concurrent -t nat -D PREROUTING -i vnet0 -j libvirt-J-vnet0' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/ebtables --concurrent -t nat -D POSTROUTING -o vnet0 -j libvirt-P-vnet0' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/ebtables --concurrent -t nat -L libvirt-J-vnet0' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/ebtables --concurrent -t nat -L libvirt-P-vnet0' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/ebtables --concurrent -t nat -F libvirt-J-vnet0' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/ebtables --concurrent -t nat -X libvirt-J-vnet0' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/ebtables --concurrent -t nat -F libvirt-P-vnet0' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/ebtables --concurrent -t nat -X libvirt-P-vnet0' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/ebtables --concurrent -t nat -F J-vnet0-mac' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/ebtables --concurrent -t nat -X J-vnet0-mac' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/ebtables --concurrent -t nat -F J-vnet0-arp-mac' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/ebtables --concurrent -t nat -X J-vnet0-arp-mac' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/iptables -w2 -w -D libvirt-out -m physdev --physdev-is-bridged --physdev-out vnet0 -g FO-vnet0' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/iptables -w2 -w -D libvirt-out -m physdev --physdev-out vnet0 -g FO-vnet0' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/iptables -w2 -w -D libvirt-in -m physdev --physdev-in vnet0 -g FI-vnet0' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/iptables -w2 -w -D libvirt-host-in -m physdev --physdev-in vnet0 -g HI-vnet0' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/iptables -w2 -w -F FO-vnet0' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/iptables -w2 -w -X FO-vnet0' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/iptables -w2 -w -F FI-vnet0' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/iptables -w2 -w -X FI-vnet0' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/iptables -w2 -w -F HI-vnet0' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/iptables -w2 -w -X HI-vnet0' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/iptables -w2 -w -E FP-vnet0 FO-vnet0' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/iptables -w2 -w -E FJ-vnet0 FI-vnet0' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/iptables -w2 -w -E HJ-vnet0 HI-vnet0' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/ip6tables -w2 -w -D libvirt-out -m physdev --physdev-is-bridged --physdev-out vnet0 -g FO-vnet0' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/ip6tables -w2 -w -D libvirt-out -m physdev --physdev-out vnet0 -g FO-vnet0' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/ip6tables -w2 -w -D libvirt-in -m physdev --physdev-in vnet0 -g FI-vnet0' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/ip6tables -w2 -w -D libvirt-host-in -m physdev --physdev-in vnet0 -g HI-vnet0' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/ip6tables -w2 -w -F FO-vnet0' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/ip6tables -w2 -w -X FO-vnet0' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/ip6tables -w2 -w -F FI-vnet0' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/ip6tables -w2 -w -X FI-vnet0' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/ip6tables -w2 -w -F HI-vnet0' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/ip6tables -w2 -w -X HI-vnet0' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/ip6tables -w2 -w -E FP-vnet0 FO-vnet0' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/ip6tables -w2 -w -E FJ-vnet0 FI-vnet0' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/ip6tables -w2 -w -E HJ-vnet0 HI-vnet0' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/ebtables --concurrent -t nat -D PREROUTING -i vnet0 -j libvirt-I-vnet0' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/ebtables --concurrent -t nat -D POSTROUTING -o vnet0 -j libvirt-O-vnet0' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/ebtables --concurrent -t nat -L libvirt-I-vnet0' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/ebtables --concurrent -t nat -L libvirt-O-vnet0' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/ebtables --concurrent -t nat -F libvirt-I-vnet0' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/ebtables --concurrent -t nat -X libvirt-I-vnet0' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/ebtables --concurrent -t nat -F libvirt-O-vnet0' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/ebtables --concurrent -t nat -X libvirt-O-vnet0' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/ebtables --concurrent -t nat -L libvirt-P-vnet0' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/ebtables --concurrent -t nat -E libvirt-P-vnet0 libvirt-O-vnet0' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/ebtables --concurrent -t nat -F I-vnet0-mac' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/ebtables --concurrent -t nat -X I-vnet0-mac' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/ebtables --concurrent -t nat -F I-vnet0-arp-mac' failed: Dec 28 23:31:35 h2 firewalld: WARNING: COMMAND_FAILED: '/usr/sbin/ebtables --concurrent -t nat -X I-vnet0-arp-mac' failed:
[root@h2 etc]# ovs-vsctl show ebb6aede-cbbc-4f4f-a88a-a9cd72b2bd23 Bridge ovirtbridge Port "ovirtport0" Interface "ovirtport0" type: internal Port ovirtbridge Interface ovirtbridge type: internal Bridge "ovsbridge0" Port "ovsbridge0" Interface "ovsbridge0" type: internal Port "eth0" Interface "eth0" Bridge br-int Port br-int Interface br-int type: internal Port "vnet0" Interface "vnet0" ovs_version: "2.6.90"
Searching through the code it appears that br-int comes from neutron-openvswitch plugin ??
[root@h2 share]# rpm -qf /usr/share/otopi/plugins/ovirt-host-deploy/openstack/neutron_openvswitch.py ovirt-host-deploy-1.6.0-0.0.master.20161215101008.gitb76ad50.el7.centos.noarch
/Sverker
Den 2016-12-28 kl. 23:24, skrev Sverker Abrahamsson:
In addition I had to add an alias to modprobe:
[root@h2 modprobe.d]# cat dummy.conf alias dummy0 dummy
Den 2016-12-28 kl. 23:03, skrev Sverker Abrahamsson:
Hi I first tried to set device name to dummy_0, but then ifup did not succeed in creating the device unless I first did 'ip link add dummy_0 type dummy' but then it would not suceed to establish the if on reboot.
Setting fake_nics = dummy0 would not work neither, but this works:
fake_nics = dummy*
The engine is now able to find the if and assign bridge ovirtmgmt to it.
However, I then run into the next issue when starting a VM:
2016-12-28 22:28:23,897 ERROR [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (ForkJoinPool-1-worker-2) [] Correlation ID: null, Call Stack: null, Custom Event ID: -1, Message: VM CentOS7 is down with error. Exit message: Cannot get interface MTU on 'br-int': No such device.
This VM has a nic on ovirtbridge, which comes from the OVN provider.
/Sverker
Den 2016-12-28 kl. 14:38, skrev Marcin Mirecki: > Sverker, > > Can you try adding a vnic named veth_* or dummy_*, > (or alternatively add the name of the vnic to > vdsm.config fake_nics), and setup the management > network using this vnic? > I suppose adding the vnic you use for connecting > to the engine to fake_nics should make it visible > to the engine, and you should be able to use it for > the setup. > > Marcin > > > > ----- Original Message ----- >> From: "Marcin Mirecki" <mmirecki@redhat.com> >> To: "Sverker Abrahamsson" <sverker@abrahamsson.com> >> Cc: "Ovirt Users" <users@ovirt.org> >> Sent: Wednesday, December 28, 2016 12:06:26 PM >> Subject: Re: [ovirt-users] Issue with OVN/OVS and mandatory >> ovirtmgmt network >> >>> I have an internal OVS bridge called ovirtbridge which has a port >>> with >>> IP address, but in the host network settings that port is not >>> visible. >> I just verified and unfortunately the virtual ports are not >> visible in engine >> to assign a network to :( >> I'm afraid that the engine is not ready for such a scenario (even >> if it >> works). >> Please give me some time to look for a solution. >> >> ----- Original Message ----- >>> From: "Sverker Abrahamsson" <sverker@abrahamsson.com> >>> To: "Marcin Mirecki" <mmirecki@redhat.com> >>> Cc: "Ovirt Users" <users@ovirt.org> >>> Sent: Wednesday, December 28, 2016 11:48:24 AM >>> Subject: Re: [ovirt-users] Issue with OVN/OVS and mandatory >>> ovirtmgmt >>> network >>> >>> Hi Marcin >>> Yes, that is my issue. I don't want to let ovirt/vdsm see eth0 nor >>> ovsbridge0 since as soon as it sees them it messes up the network >>> config >>> so that the host will be unreachable. >>> >>> I have an internal OVS bridge called ovirtbridge which has a port >>> with >>> IP address, but in the host network settings that port is not >>> visible. >>> It doesn't help to name it ovirtmgmt. >>> >>> The engine is able to communicate with the host on the ip it has >>> been >>> given, it's just that it believes that it HAS to have a ovirtmgmt >>> network which can't be on OVN. >>> >>> /Sverker >>> >>> >>> Den 2016-12-28 kl. 10:45, skrev Marcin Mirecki: >>>> Hi Sverker, >>>> >>>> The management network is mandatory on each host. It's used by the >>>> engine to communicate with the host. >>>> Looking at your description and the exception it looks like it is >>>> missing. >>>> The error is caused by not having any network for the host >>>> (network list retrieved in >>>> InterfaceDaoImpl.getHostNetworksByCluster - >>>> which >>>> gets all the networks on nics for a host from vds_interface >>>> table in the >>>> DB). >>>> >>>> Could you maybe create a virtual nic connected to ovsbridge0 (as I >>>> understand you >>>> have no physical nic available) and use this for the management >>>> network? >>>> >>>>> I then create a bridge for use with ovirt, with a private address. >>>> I'm not quite sure I understand. Is this yet another bridge >>>> connected to >>>> ovsbridge0? >>>> You could also attach the vnic for the management network here >>>> if need >>>> be. >>>> >>>> Please keep in mind that OVN has no use in setting up the >>>> management >>>> network. >>>> The OVN provider can only handle external networks, which can >>>> not be used >>>> for a >>>> management network. >>>> >>>> Marcin >>>> >>>> >>>> ----- Original Message ----- >>>>> From: "Sverker Abrahamsson" <sverker@abrahamsson.com> >>>>> To: users@ovirt.org >>>>> Sent: Wednesday, December 28, 2016 12:39:59 AM >>>>> Subject: [ovirt-users] Issue with OVN/OVS and mandatory ovirtmgmt >>>>> network >>>>> >>>>> >>>>> >>>>> Hi >>>>> For long time I've been looking for proper support in ovirt for >>>>> Open >>>>> vSwitch >>>>> so I'm happy that it is moving in the right direction. However, >>>>> there >>>>> seems >>>>> to still be a dependency on a ovirtmgmt bridge and I'm unable >>>>> to move >>>>> that >>>>> to the OVN provider. >>>>> >>>>> The hosting center where I rent hw instances has a bit special >>>>> network >>>>> setup, >>>>> so I have one physical network port with a /32 netmask and >>>>> point-to-point >>>>> config to router. The physical port I connect to a ovs bridge >>>>> which has >>>>> the >>>>> public ip. Since ovirt always messes up the network config when >>>>> I've >>>>> tried >>>>> to let it have access to the network config for the physical >>>>> port, I've >>>>> set >>>>> eht0 and ovsbridge0 as hidden in vdsm.conf. >>>>> >>>>> >>>>> I then create a bridge for use with ovirt, with a private >>>>> address. With >>>>> the >>>>> OVN provider I am now able to import these into the engine and >>>>> it looks >>>>> good. When creating a VM I can select that it will have a vNic >>>>> on my OVS >>>>> bridge. >>>>> >>>>> However, I can't start the VM as an exception is thrown in the >>>>> log: >>>>> >>>>> 2016-12-28 00:13:33,350 ERROR >>>>> [org.ovirt.engine.core.bll.RunVmCommand] >>>>> (default task-5) [3c882d53] Error during ValidateFailure.: >>>>> java.lang.NullPointerException >>>>> at >>>>> org.ovirt.engine.core.bll.scheduling.policyunits.NetworkPolicyUnit.validateRequiredNetworksAvailable(NetworkPolicyUnit.java:140) >>>>> >>>>> [bll.jar:] >>>>> at >>>>> org.ovirt.engine.core.bll.scheduling.policyunits.NetworkPolicyUnit.filter(NetworkPolicyUnit.java:69) >>>>> >>>>> [bll.jar:] >>>>> at >>>>> org.ovirt.engine.core.bll.scheduling.SchedulingManager.runInternalFilters(SchedulingManager.java:597) >>>>> >>>>> [bll.jar:] >>>>> at >>>>> org.ovirt.engine.core.bll.scheduling.SchedulingManager.runFilters(SchedulingManager.java:564) >>>>> >>>>> [bll.jar:] >>>>> at >>>>> org.ovirt.engine.core.bll.scheduling.SchedulingManager.canSchedule(SchedulingManager.java:494) >>>>> >>>>> [bll.jar:] >>>>> at >>>>> org.ovirt.engine.core.bll.validator.RunVmValidator.canRunVm(RunVmValidator.java:133) >>>>> >>>>> [bll.jar:] >>>>> at >>>>> org.ovirt.engine.core.bll.RunVmCommand.validate(RunVmCommand.java:940) >>>>> >>>>> [bll.jar:] >>>>> at >>>>> org.ovirt.engine.core.bll.CommandBase.internalValidate(CommandBase.java:886) >>>>> >>>>> [bll.jar:] >>>>> at >>>>> org.ovirt.engine.core.bll.CommandBase.validateOnly(CommandBase.java:366) >>>>> >>>>> [bll.jar:] >>>>> at >>>>> org.ovirt.engine.core.bll.PrevalidatingMultipleActionsRunner.canRunActions(PrevalidatingMultipleActionsRunner.java:113) >>>>> >>>>> [bll.jar:] >>>>> at >>>>> org.ovirt.engine.core.bll.PrevalidatingMultipleActionsRunner.invokeCommands(PrevalidatingMultipleActionsRunner.java:99) >>>>> >>>>> [bll.jar:] >>>>> at >>>>> org.ovirt.engine.core.bll.PrevalidatingMultipleActionsRunner.execute(PrevalidatingMultipleActionsRunner.java:76) >>>>> >>>>> [bll.jar:] >>>>> at >>>>> org.ovirt.engine.core.bll.Backend.runMultipleActionsImpl(Backend.java:613) >>>>> >>>>> [bll.jar:] >>>>> at >>>>> org.ovirt.engine.core.bll.Backend.runMultipleActions(Backend.java:583) >>>>> >>>>> [bll.jar:] >>>>> >>>>> >>>>> Looking at that section of code where the exception is thrown, >>>>> I see >>>>> that >>>>> it >>>>> iterates over host networks to find required networks, which I >>>>> assume is >>>>> ovirtmgmt. In the host network setup dialog I don't see any >>>>> networks at >>>>> all >>>>> but it lists ovirtmgmt as required. It also list the OVN >>>>> networks but >>>>> these >>>>> can't be statically assigned as they are added dynamically when >>>>> needed, >>>>> which is fine. >>>>> >>>>> I believe that I either need to remove ovirtmgmt network or >>>>> configure >>>>> that >>>>> it >>>>> is provided by the OVN provider, but neither is possible. >>>>> Preferably it >>>>> shouldn't be hardcoded which network is management and >>>>> mandatory but be >>>>> possible to configure. >>>>> >>>>> /Sverker >>>>> Den 2016-12-27 kl. 17:10, skrev Marcin Mirecki: >>>>> >>>>> >> _______________________________________________ >> Users mailing list >> Users@ovirt.org >> http://lists.ovirt.org/mailman/listinfo/users >> _______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
_______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users