Sure
I've attached it for easier reference.
On Wed, Sep 30, 2020 at 2:21 PM Dominik Holler <dholler(a)redhat.com> wrote:
On Wed, Sep 30, 2020 at 1:16 PM Konstantinos Betsis <k.betsis(a)gmail.com>
wrote:
> Hi Dominik
>
> The DC01-node02 was formatted and reinstalled and then attached to ovirt
> environment.
> Unfortunately we exhibit the same issue.
> The new DC01-node02 tries to establish geneve tunnels to his own IP.
>
> [root@dc01-node02 ~]# ovs-vsctl show
> eff2663e-cb10-41b0-93ba-605bb5c7bd78
> Bridge br-int
> fail_mode: secure
> Port "ovn-95ccb0-0"
> Interface "ovn-95ccb0-0"
> type: geneve
> options: {csum="true", key=flow,
> remote_ip="dc01-node01_IP"}
> Port "ovn-be3abc-0"
> Interface "ovn-be3abc-0"
> type: geneve
> options: {csum="true", key=flow,
> remote_ip="dc01-node02_IP"}
> Port "ovn-c4b238-0"
> Interface "ovn-c4b238-0"
> type: geneve
> options: {csum="true", key=flow,
> remote_ip="dc02-node01_IP"}
> Port br-int
> Interface br-int
> type: internal
> ovs_version: "2.11.0"
>
>
> Is there a way to fix this on the Ovirt engine since this is where the
> information resides?
> Something is broken there.
>
>
I suspect that there is an inconsistency in the OVN SB DB.
Is there a way to share your /var/lib/openvswitch/ovnsb_db.db with us?
> Thank you
> Best Regards
> Konstantinos Betsis
>
>
> On Wed, Sep 16, 2020, 13:25 Dominik Holler <dholler(a)redhat.com> wrote:
>
>>
>>
>> On Wed, Sep 16, 2020 at 12:15 PM Konstantinos Betsis <k.betsis(a)gmail.com>
>> wrote:
>>
>>> I have a better solution.
>>> I am currently migrating all VMs over to dc01-node01 and then i'll
>>> format it as to fix the partitioning as well.
>>>
>>>
>>
>>> In theory the ovs sbdb will be fixed once it is re-installed....
>>>
>>> If not we can then check if there is a stale entry in the ovirt host
>>> where the sb db is managed.
>>>
>>>
>> Maybe you could ensure that there is no entry anymore during dc01-host02
>> is reinstalling, before the host is added to oVirt again?
>>
>>
>>> Do you agree with this?
>>>
>>>
>> Sounds good, but OVN should be not the reason to reinstall.
>>
>>
>>> On Wed, Sep 16, 2020 at 1:00 PM Dominik Holler <dholler(a)redhat.com>
>>> wrote:
>>>
>>>>
>>>>
>>>> Maybe because of a duplicated entry in the ovn sb db?
>>>> Can you please stop the ovn-ctrontroller on this host, remove the host
>>>> from the ovn sb db, ensure it is gone and restart the ovn-controller on
the
>>>> host?
>>>>
>>>>
>>>> On Wed, Sep 16, 2020 at 11:55 AM Konstantinos Betsis <
>>>> k.betsis(a)gmail.com> wrote:
>>>>
>>>>> Hi Dominik
>>>>>
>>>>> Just saw the below on host dc01-host02
>>>>>
>>>>> ovs-vsctl show
>>>>> f3b13557-dfb4-45a4-b6af-c995ccf68720
>>>>> Bridge br-int
>>>>> Port "ovn-95ccb0-0"
>>>>> Interface "ovn-95ccb0-0"
>>>>> type: geneve
>>>>> options: {csum="true", key=flow,
>>>>> remote_ip="dc01-host01"}
>>>>> Port "vnet10"
>>>>> Interface "vnet10"
>>>>> Port "vnet11"
>>>>> Interface "vnet11"
>>>>> Port "vnet0"
>>>>> Interface "vnet0"
>>>>> Port "vnet9"
>>>>> Interface "vnet9"
>>>>> Port "vnet8"
>>>>> Interface "vnet8"
>>>>> Port br-int
>>>>> Interface br-int
>>>>> type: internal
>>>>> Port "vnet12"
>>>>> Interface "vnet12"
>>>>> Port "ovn-be3abc-0"
>>>>> Interface "ovn-be3abc-0"
>>>>> type: geneve
>>>>> options: {csum="true", key=flow,
>>>>> remote_ip="dc01-host02"}
>>>>> Port "vnet7"
>>>>> Interface "vnet7"
>>>>> Port "ovn-c4b238-0"
>>>>> Interface "ovn-c4b238-0"
>>>>> type: geneve
>>>>> options: {csum="true", key=flow,
>>>>> remote_ip="dc02-host01"}
>>>>> Port "vnet6"
>>>>> Interface "vnet6"
>>>>> ovs_version: "2.11.0"
>>>>>
>>>>>
>>>>> Why would this node establish a geneve tunnel to himself?
>>>>> Other nodes do not exhibit this behavior.
>>>>>
>>>>> On Wed, Sep 16, 2020 at 12:21 PM Konstantinos Betsis <
>>>>> k.betsis(a)gmail.com> wrote:
>>>>>
>>>>>> Hi Dominik
>>>>>>
>>>>>> Below is the output of the ovs-vsctl list interface
>>>>>>
>>>>>> _uuid : bdaf92c1-4389-4ddf-aab0-93975076ebb2
>>>>>> admin_state : up
>>>>>> bfd : {}
>>>>>> bfd_status : {}
>>>>>> cfm_fault : []
>>>>>> cfm_fault_status : []
>>>>>> cfm_flap_count : []
>>>>>> cfm_health : []
>>>>>> cfm_mpid : []
>>>>>> cfm_remote_mpids : []
>>>>>> cfm_remote_opstate : []
>>>>>> duplex : full
>>>>>> error : []
>>>>>> external_ids :
{attached-mac="56:6f:77:61:00:02",
>>>>>> iface-id="5d03a7a5-82a1-40f9-b50c-353a26167fa3",
iface-status=active,
>>>>>> vm-id="e45b7b34-24f2-41ed-b95e-cd1ad532e8d3"}
>>>>>> ifindex : 34
>>>>>> ingress_policing_burst: 0
>>>>>> ingress_policing_rate: 0
>>>>>> lacp_current : []
>>>>>> link_resets : 1
>>>>>> link_speed : 10000000
>>>>>> link_state : up
>>>>>> lldp : {}
>>>>>> mac : []
>>>>>> mac_in_use : "fe:6f:77:61:00:02"
>>>>>> mtu : 1442
>>>>>> mtu_request : []
>>>>>> name : "vnet6"
>>>>>> ofport : 2
>>>>>> ofport_request : []
>>>>>> options : {}
>>>>>> other_config : {}
>>>>>> statistics : {collisions=0, rx_bytes=10828495,
>>>>>> rx_crc_err=0, rx_dropped=0, rx_errors=0, rx_frame_err=0,
rx_over_err=0,
>>>>>> rx_packets=117713, tx_bytes=20771797, tx_dropped=0, tx_errors=0,
>>>>>> tx_packets=106954}
>>>>>> status : {driver_name=tun,
driver_version="1.6",
>>>>>> firmware_version=""}
>>>>>> type : ""
>>>>>>
>>>>>> _uuid : bad80911-3993-4085-a0b0-962b6c9156cd
>>>>>> admin_state : up
>>>>>> bfd : {}
>>>>>> bfd_status : {}
>>>>>> cfm_fault : []
>>>>>> cfm_fault_status : []
>>>>>> cfm_flap_count : []
>>>>>> cfm_health : []
>>>>>> cfm_mpid : []
>>>>>> cfm_remote_mpids : []
>>>>>> cfm_remote_opstate : []
>>>>>> duplex : []
>>>>>> error : []
>>>>>> external_ids : {}
>>>>>> ifindex : 39
>>>>>> ingress_policing_burst: 0
>>>>>> ingress_policing_rate: 0
>>>>>> lacp_current : []
>>>>>> link_resets : 0
>>>>>> link_speed : []
>>>>>> link_state : up
>>>>>> lldp : {}
>>>>>> mac : []
>>>>>> mac_in_use : "fe:37:52:c4:cb:03"
>>>>>> mtu : []
>>>>>> mtu_request : []
>>>>>> name : "ovn-c4b238-0"
>>>>>> ofport : 7
>>>>>> ofport_request : []
>>>>>> options : {csum="true", key=flow,
>>>>>> remote_ip="192.168.121.164"}
>>>>>> other_config : {}
>>>>>> statistics : {rx_bytes=0, rx_packets=0, tx_bytes=0,
>>>>>> tx_packets=0}
>>>>>> status :
{tunnel_egress_iface="ovirtmgmt-ams03",
>>>>>> tunnel_egress_iface_carrier=up}
>>>>>> type : geneve
>>>>>>
>>>>>> _uuid : 8e7705d1-0b9d-4e30-8277-c339e7e1c27a
>>>>>> admin_state : up
>>>>>> bfd : {}
>>>>>> bfd_status : {}
>>>>>> cfm_fault : []
>>>>>> cfm_fault_status : []
>>>>>> cfm_flap_count : []
>>>>>> cfm_health : []
>>>>>> cfm_mpid : []
>>>>>> cfm_remote_mpids : []
>>>>>> cfm_remote_opstate : []
>>>>>> duplex : full
>>>>>> error : []
>>>>>> external_ids :
{attached-mac="56:6f:77:61:00:0d",
>>>>>> iface-id="b7ff5f2b-4bb4-4250-8ad8-8a7e19d2b4c7",
iface-status=active,
>>>>>> vm-id="8d73f333-bca4-4b32-9b87-2e7ee07eda84"}
>>>>>> ifindex : 28
>>>>>> ingress_policing_burst: 0
>>>>>> ingress_policing_rate: 0
>>>>>> lacp_current : []
>>>>>> link_resets : 1
>>>>>> link_speed : 10000000
>>>>>> link_state : up
>>>>>> lldp : {}
>>>>>> mac : []
>>>>>> mac_in_use : "fe:6f:77:61:00:0d"
>>>>>> mtu : 1442
>>>>>> mtu_request : []
>>>>>> name : "vnet0"
>>>>>> ofport : 1
>>>>>> ofport_request : []
>>>>>> options : {}
>>>>>> other_config : {}
>>>>>> statistics : {collisions=0, rx_bytes=20609787,
>>>>>> rx_crc_err=0, rx_dropped=0, rx_errors=0, rx_frame_err=0,
rx_over_err=0,
>>>>>> rx_packets=104535, tx_bytes=10830007, tx_dropped=0, tx_errors=0,
>>>>>> tx_packets=117735}
>>>>>> status : {driver_name=tun,
driver_version="1.6",
>>>>>> firmware_version=""}
>>>>>> type : ""
>>>>>>
>>>>>> _uuid : 86dcc68a-63e4-4445-9373-81c1f4502c17
>>>>>> admin_state : up
>>>>>> bfd : {}
>>>>>> bfd_status : {}
>>>>>> cfm_fault : []
>>>>>> cfm_fault_status : []
>>>>>> cfm_flap_count : []
>>>>>> cfm_health : []
>>>>>> cfm_mpid : []
>>>>>> cfm_remote_mpids : []
>>>>>> cfm_remote_opstate : []
>>>>>> duplex : full
>>>>>> error : []
>>>>>> external_ids :
{attached-mac="56:6f:77:61:00:10",
>>>>>> iface-id="4e8d5636-4110-41b2-906d-f9b04c2e62cd",
iface-status=active,
>>>>>> vm-id="9a002a9b-5f09-4def-a531-d50ff683470b"}
>>>>>> ifindex : 40
>>>>>> ingress_policing_burst: 0
>>>>>> ingress_policing_rate: 0
>>>>>> lacp_current : []
>>>>>> link_resets : 1
>>>>>> link_speed : 10000000
>>>>>> link_state : up
>>>>>> lldp : {}
>>>>>> mac : []
>>>>>> mac_in_use : "fe:6f:77:61:00:10"
>>>>>> mtu : 1442
>>>>>> mtu_request : []
>>>>>> name : "vnet11"
>>>>>> ofport : 10
>>>>>> ofport_request : []
>>>>>> options : {}
>>>>>> other_config : {}
>>>>>> statistics : {collisions=0, rx_bytes=3311352,
rx_crc_err=0,
>>>>>> rx_dropped=0, rx_errors=0, rx_frame_err=0, rx_over_err=0,
rx_packets=51012,
>>>>>> tx_bytes=5514116, tx_dropped=0, tx_errors=0, tx_packets=103456}
>>>>>> status : {driver_name=tun,
driver_version="1.6",
>>>>>> firmware_version=""}
>>>>>> type : ""
>>>>>>
>>>>>> _uuid : e8d5e4a2-b9a0-4146-8d98-34713cb443de
>>>>>> admin_state : up
>>>>>> bfd : {}
>>>>>> bfd_status : {}
>>>>>> cfm_fault : []
>>>>>> cfm_fault_status : []
>>>>>> cfm_flap_count : []
>>>>>> cfm_health : []
>>>>>> cfm_mpid : []
>>>>>> cfm_remote_mpids : []
>>>>>> cfm_remote_opstate : []
>>>>>> duplex : full
>>>>>> error : []
>>>>>> external_ids :
{attached-mac="56:6f:77:61:00:15",
>>>>>> iface-id="b88de6e4-6d77-4e42-b734-4cc676728910",
iface-status=active,
>>>>>> vm-id="e45b7b34-24f2-41ed-b95e-cd1ad532e8d3"}
>>>>>> ifindex : 37
>>>>>> ingress_policing_burst: 0
>>>>>> ingress_policing_rate: 0
>>>>>> lacp_current : []
>>>>>> link_resets : 1
>>>>>> link_speed : 10000000
>>>>>> link_state : up
>>>>>> lldp : {}
>>>>>> mac : []
>>>>>> mac_in_use : "fe:6f:77:61:00:15"
>>>>>> mtu : 1442
>>>>>> mtu_request : []
>>>>>> name : "vnet9"
>>>>>> ofport : 5
>>>>>> ofport_request : []
>>>>>> options : {}
>>>>>> other_config : {}
>>>>>> statistics : {collisions=0, rx_bytes=180, rx_crc_err=0,
>>>>>> rx_dropped=0, rx_errors=0, rx_frame_err=0, rx_over_err=0,
rx_packets=2,
>>>>>> tx_bytes=4500, tx_dropped=0, tx_errors=0, tx_packets=74}
>>>>>> status : {driver_name=tun,
driver_version="1.6",
>>>>>> firmware_version=""}
>>>>>> type : ""
>>>>>>
>>>>>> _uuid : 6a2974b3-cd72-4688-a630-0a7e9c779b21
>>>>>> admin_state : up
>>>>>> bfd : {}
>>>>>> bfd_status : {}
>>>>>> cfm_fault : []
>>>>>> cfm_fault_status : []
>>>>>> cfm_flap_count : []
>>>>>> cfm_health : []
>>>>>> cfm_mpid : []
>>>>>> cfm_remote_mpids : []
>>>>>> cfm_remote_opstate : []
>>>>>> duplex : full
>>>>>> error : []
>>>>>> external_ids :
{attached-mac="56:6f:77:61:00:17",
>>>>>> iface-id="64681036-26e2-41d7-b73f-ab5302610145",
iface-status=active,
>>>>>> vm-id="bf0dc78c-dad5-41a0-914c-ae0da0f9a388"}
>>>>>> ifindex : 41
>>>>>> ingress_policing_burst: 0
>>>>>> ingress_policing_rate: 0
>>>>>> lacp_current : []
>>>>>> link_resets : 1
>>>>>> link_speed : 10000000
>>>>>> link_state : up
>>>>>> lldp : {}
>>>>>> mac : []
>>>>>> mac_in_use : "fe:6f:77:61:00:17"
>>>>>> mtu : 1442
>>>>>> mtu_request : []
>>>>>> name : "vnet12"
>>>>>> ofport : 11
>>>>>> ofport_request : []
>>>>>> options : {}
>>>>>> other_config : {}
>>>>>> statistics : {collisions=0, rx_bytes=5513640,
rx_crc_err=0,
>>>>>> rx_dropped=0, rx_errors=0, rx_frame_err=0, rx_over_err=0,
>>>>>> rx_packets=103450, tx_bytes=3311868, tx_dropped=0, tx_errors=0,
>>>>>> tx_packets=51018}
>>>>>> status : {driver_name=tun,
driver_version="1.6",
>>>>>> firmware_version=""}
>>>>>> type : ""
>>>>>>
>>>>>> _uuid : 44498e54-f122-41a0-a41a-7a88ba2dba9b
>>>>>> admin_state : down
>>>>>> bfd : {}
>>>>>> bfd_status : {}
>>>>>> cfm_fault : []
>>>>>> cfm_fault_status : []
>>>>>> cfm_flap_count : []
>>>>>> cfm_health : []
>>>>>> cfm_mpid : []
>>>>>> cfm_remote_mpids : []
>>>>>> cfm_remote_opstate : []
>>>>>> duplex : []
>>>>>> error : []
>>>>>> external_ids : {}
>>>>>> ifindex : 7
>>>>>> ingress_policing_burst: 0
>>>>>> ingress_policing_rate: 0
>>>>>> lacp_current : []
>>>>>> link_resets : 0
>>>>>> link_speed : []
>>>>>> link_state : down
>>>>>> lldp : {}
>>>>>> mac : []
>>>>>> mac_in_use : "32:0a:69:67:07:4f"
>>>>>> mtu : 1442
>>>>>> mtu_request : []
>>>>>> name : br-int
>>>>>> ofport : 65534
>>>>>> ofport_request : []
>>>>>> options : {}
>>>>>> other_config : {}
>>>>>> statistics : {collisions=0, rx_bytes=0, rx_crc_err=0,
>>>>>> rx_dropped=326, rx_errors=0, rx_frame_err=0, rx_over_err=0,
rx_packets=0,
>>>>>> tx_bytes=0, tx_dropped=0, tx_errors=0, tx_packets=0}
>>>>>> status : {driver_name=openvswitch}
>>>>>> type : internal
>>>>>>
>>>>>> _uuid : e2114584-8ceb-43d6-817b-e457738ead8a
>>>>>> admin_state : up
>>>>>> bfd : {}
>>>>>> bfd_status : {}
>>>>>> cfm_fault : []
>>>>>> cfm_fault_status : []
>>>>>> cfm_flap_count : []
>>>>>> cfm_health : []
>>>>>> cfm_mpid : []
>>>>>> cfm_remote_mpids : []
>>>>>> cfm_remote_opstate : []
>>>>>> duplex : full
>>>>>> error : []
>>>>>> external_ids :
{attached-mac="56:6f:77:61:00:03",
>>>>>> iface-id="16162721-c815-4cd8-ab57-f22e6e482c7f",
iface-status=active,
>>>>>> vm-id="e45b7b34-24f2-41ed-b95e-cd1ad532e8d3"}
>>>>>> ifindex : 35
>>>>>> ingress_policing_burst: 0
>>>>>> ingress_policing_rate: 0
>>>>>> lacp_current : []
>>>>>> link_resets : 1
>>>>>> link_speed : 10000000
>>>>>> link_state : up
>>>>>> lldp : {}
>>>>>> mac : []
>>>>>> mac_in_use : "fe:6f:77:61:00:03"
>>>>>> mtu : 1442
>>>>>> mtu_request : []
>>>>>> name : "vnet7"
>>>>>> ofport : 3
>>>>>> ofport_request : []
>>>>>> options : {}
>>>>>> other_config : {}
>>>>>> statistics : {collisions=0, rx_bytes=180, rx_crc_err=0,
>>>>>> rx_dropped=0, rx_errors=0, rx_frame_err=0, rx_over_err=0,
rx_packets=2,
>>>>>> tx_bytes=4730, tx_dropped=0, tx_errors=0, tx_packets=77}
>>>>>> status : {driver_name=tun,
driver_version="1.6",
>>>>>> firmware_version=""}
>>>>>> type : ""
>>>>>>
>>>>>> _uuid : ee16943e-d145-4080-893f-464098a6388f
>>>>>> admin_state : up
>>>>>> bfd : {}
>>>>>> bfd_status : {}
>>>>>> cfm_fault : []
>>>>>> cfm_fault_status : []
>>>>>> cfm_flap_count : []
>>>>>> cfm_health : []
>>>>>> cfm_mpid : []
>>>>>> cfm_remote_mpids : []
>>>>>> cfm_remote_opstate : []
>>>>>> duplex : []
>>>>>> error : []
>>>>>> external_ids : {}
>>>>>> ifindex : 39
>>>>>> ingress_policing_burst: 0
>>>>>> ingress_policing_rate: 0
>>>>>> lacp_current : []
>>>>>> link_resets : 0
>>>>>> link_speed : []
>>>>>> link_state : up
>>>>>> lldp : {}
>>>>>> mac : []
>>>>>> mac_in_use : "1e:50:3f:a8:42:d1"
>>>>>> mtu : []
>>>>>> mtu_request : []
>>>>>> name : "ovn-be3abc-0"
>>>>>> ofport : 8
>>>>>> ofport_request : []
>>>>>> options : {csum="true", key=flow,
>>>>>> remote_ip="DC01-host02"}
>>>>>> other_config : {}
>>>>>> statistics : {rx_bytes=0, rx_packets=0, tx_bytes=0,
>>>>>> tx_packets=0}
>>>>>> status :
{tunnel_egress_iface="ovirtmgmt-ams03",
>>>>>> tunnel_egress_iface_carrier=up}
>>>>>> type : geneve
>>>>>>
>>>>>> _uuid : 86a229be-373e-4c43-b2f1-6190523ed73a
>>>>>> admin_state : up
>>>>>> bfd : {}
>>>>>> bfd_status : {}
>>>>>> cfm_fault : []
>>>>>> cfm_fault_status : []
>>>>>> cfm_flap_count : []
>>>>>> cfm_health : []
>>>>>> cfm_mpid : []
>>>>>> cfm_remote_mpids : []
>>>>>> cfm_remote_opstate : []
>>>>>> duplex : full
>>>>>> error : []
>>>>>> external_ids :
{attached-mac="56:6f:77:61:00:1c",
>>>>>> iface-id="12d829c3-64eb-44bc-a0bd-d7219991f35f",
iface-status=active,
>>>>>> vm-id="e45b7b34-24f2-41ed-b95e-cd1ad532e8d3"}
>>>>>> ifindex : 38
>>>>>> ingress_policing_burst: 0
>>>>>> ingress_policing_rate: 0
>>>>>> lacp_current : []
>>>>>> link_resets : 1
>>>>>> link_speed : 10000000
>>>>>> link_state : up
>>>>>> lldp : {}
>>>>>> mac : []
>>>>>> mac_in_use : "fe:6f:77:61:00:1c"
>>>>>> mtu : 1442
>>>>>> mtu_request : []
>>>>>> name : "vnet10"
>>>>>> ofport : 6
>>>>>> ofport_request : []
>>>>>> options : {}
>>>>>> other_config : {}
>>>>>> statistics : {collisions=0, rx_bytes=117912,
rx_crc_err=0,
>>>>>> rx_dropped=0, rx_errors=0, rx_frame_err=0, rx_over_err=0,
rx_packets=2195,
>>>>>> tx_bytes=4204, tx_dropped=0, tx_errors=0, tx_packets=66}
>>>>>> status : {driver_name=tun,
driver_version="1.6",
>>>>>> firmware_version=""}
>>>>>> type : ""
>>>>>>
>>>>>> _uuid : fa4b8d96-bffe-4b56-930e-0e7fcc5f68ac
>>>>>> admin_state : up
>>>>>> bfd : {}
>>>>>> bfd_status : {}
>>>>>> cfm_fault : []
>>>>>> cfm_fault_status : []
>>>>>> cfm_flap_count : []
>>>>>> cfm_health : []
>>>>>> cfm_mpid : []
>>>>>> cfm_remote_mpids : []
>>>>>> cfm_remote_opstate : []
>>>>>> duplex : []
>>>>>> error : []
>>>>>> external_ids : {}
>>>>>> ifindex : 39
>>>>>> ingress_policing_burst: 0
>>>>>> ingress_policing_rate: 0
>>>>>> lacp_current : []
>>>>>> link_resets : 0
>>>>>> link_speed : []
>>>>>> link_state : up
>>>>>> lldp : {}
>>>>>> mac : []
>>>>>> mac_in_use : "7a:28:24:eb:ec:d2"
>>>>>> mtu : []
>>>>>> mtu_request : []
>>>>>> name : "ovn-95ccb0-0"
>>>>>> ofport : 9
>>>>>> ofport_request : []
>>>>>> options : {csum="true", key=flow,
>>>>>> remote_ip="DC01-host01"}
>>>>>> other_config : {}
>>>>>> statistics : {rx_bytes=0, rx_packets=0,
tx_bytes=12840478,
>>>>>> tx_packets=224029}
>>>>>> status :
{tunnel_egress_iface="ovirtmgmt-ams03",
>>>>>> tunnel_egress_iface_carrier=up}
>>>>>> type : geneve
>>>>>>
>>>>>> _uuid : 5e3df5c7-958c-491d-8d41-0ae83c613f1d
>>>>>> admin_state : up
>>>>>> bfd : {}
>>>>>> bfd_status : {}
>>>>>> cfm_fault : []
>>>>>> cfm_fault_status : []
>>>>>> cfm_flap_count : []
>>>>>> cfm_health : []
>>>>>> cfm_mpid : []
>>>>>> cfm_remote_mpids : []
>>>>>> cfm_remote_opstate : []
>>>>>> duplex : full
>>>>>> error : []
>>>>>> external_ids :
{attached-mac="56:6f:77:61:00:06",
>>>>>> iface-id="9a6cc189-0934-4468-97ae-09f90fa4598d",
iface-status=active,
>>>>>> vm-id="e45b7b34-24f2-41ed-b95e-cd1ad532e8d3"}
>>>>>> ifindex : 36
>>>>>> ingress_policing_burst: 0
>>>>>> ingress_policing_rate: 0
>>>>>> lacp_current : []
>>>>>> link_resets : 1
>>>>>> link_speed : 10000000
>>>>>> link_state : up
>>>>>> lldp : {}
>>>>>> mac : []
>>>>>> mac_in_use : "fe:6f:77:61:00:06"
>>>>>> mtu : 1442
>>>>>> mtu_request : []
>>>>>> name : "vnet8"
>>>>>> ofport : 4
>>>>>> ofport_request : []
>>>>>> options : {}
>>>>>> other_config : {}
>>>>>> statistics : {collisions=0, rx_bytes=180, rx_crc_err=0,
>>>>>> rx_dropped=0, rx_errors=0, rx_frame_err=0, rx_over_err=0,
rx_packets=2,
>>>>>> tx_bytes=8829812, tx_dropped=0, tx_errors=0, tx_packets=154540}
>>>>>> status : {driver_name=tun,
driver_version="1.6",
>>>>>> firmware_version=""}
>>>>>> type : ""
>>>>>>
>>>>>>
>>>>>> I've identified which VMs have these MAC addresses but i do
not see
>>>>>> any "conflict" with any other VM's MAC address.
>>>>>>
>>>>>> I really do not understand why these will create a conflict.
>>>>>>
>>>>>> On Wed, Sep 16, 2020 at 12:06 PM Dominik Holler
<dholler(a)redhat.com>
>>>>>> wrote:
>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> On Tue, Sep 15, 2020 at 6:53 PM Konstantinos Betsis <
>>>>>>> k.betsis(a)gmail.com> wrote:
>>>>>>>
>>>>>>>> So a new test-net was created under DC01 and was depicted
in the
>>>>>>>> networks tab under both DC01 and DC02.
>>>>>>>> I believe for some reason networks are duplicated in DCs,
maybe
>>>>>>>> for future use??? Don't know.
>>>>>>>> If one tries to delete the network from the other DC it
gets an
>>>>>>>> error, while if deleted from the once initially created
it gets deleted
>>>>>>>> from both.
>>>>>>>>
>>>>>>>>
>>>>>>> In oVirt a logical network is an entity in a data center. If
the
>>>>>>> automatic synchronization is enabled on the
ovirt-provider-ovn entity in
>>>>>>> oVirt Engine, the OVN networks are reflected to all data
centers. If you do
>>>>>>> not like this, you can disable the automatic synchronization
of the
>>>>>>> ovirt-provider-ovn in Admin Portal.
>>>>>>>
>>>>>>>
>>>>>>>> From the DC01-node02 i get the following errors:
>>>>>>>>
>>>>>>>> 2020-09-15T16:48:49.904Z|22748|main|INFO|OVNSB commit
failed,
>>>>>>>> force recompute next time.
>>>>>>>> 2020-09-15T16:48:49.905Z|22749|binding|INFO|Claiming
lport
>>>>>>>> 9a6cc189-0934-4468-97ae-09f90fa4598d for this chassis.
>>>>>>>>
2020-09-15T16:48:49.905Z|22750|binding|INFO|9a6cc189-0934-4468-97ae-09f90fa4598d:
>>>>>>>> Claiming 56:6f:77:61:00:06
>>>>>>>> 2020-09-15T16:48:49.905Z|22751|binding|INFO|Claiming
lport
>>>>>>>> 16162721-c815-4cd8-ab57-f22e6e482c7f for this chassis.
>>>>>>>>
2020-09-15T16:48:49.905Z|22752|binding|INFO|16162721-c815-4cd8-ab57-f22e6e482c7f:
>>>>>>>> Claiming 56:6f:77:61:00:03
>>>>>>>> 2020-09-15T16:48:49.905Z|22753|binding|INFO|Claiming
lport
>>>>>>>> b88de6e4-6d77-4e42-b734-4cc676728910 for this chassis.
>>>>>>>>
2020-09-15T16:48:49.905Z|22754|binding|INFO|b88de6e4-6d77-4e42-b734-4cc676728910:
>>>>>>>> Claiming 56:6f:77:61:00:15
>>>>>>>> 2020-09-15T16:48:49.905Z|22755|binding|INFO|Claiming
lport
>>>>>>>> b7ff5f2b-4bb4-4250-8ad8-8a7e19d2b4c7 for this chassis.
>>>>>>>>
2020-09-15T16:48:49.905Z|22756|binding|INFO|b7ff5f2b-4bb4-4250-8ad8-8a7e19d2b4c7:
>>>>>>>> Claiming 56:6f:77:61:00:0d
>>>>>>>> 2020-09-15T16:48:49.905Z|22757|binding|INFO|Claiming
lport
>>>>>>>> 5d03a7a5-82a1-40f9-b50c-353a26167fa3 for this chassis.
>>>>>>>>
2020-09-15T16:48:49.905Z|22758|binding|INFO|5d03a7a5-82a1-40f9-b50c-353a26167fa3:
>>>>>>>> Claiming 56:6f:77:61:00:02
>>>>>>>> 2020-09-15T16:48:49.905Z|22759|binding|INFO|Claiming
lport
>>>>>>>> 12d829c3-64eb-44bc-a0bd-d7219991f35f for this chassis.
>>>>>>>>
2020-09-15T16:48:49.905Z|22760|binding|INFO|12d829c3-64eb-44bc-a0bd-d7219991f35f:
>>>>>>>> Claiming 56:6f:77:61:00:1c
>>>>>>>> 2020-09-15T16:48:49.959Z|22761|main|INFO|OVNSB commit
failed,
>>>>>>>> force recompute next time.
>>>>>>>> 2020-09-15T16:48:49.960Z|22762|binding|INFO|Claiming
lport
>>>>>>>> 9a6cc189-0934-4468-97ae-09f90fa4598d for this chassis.
>>>>>>>>
2020-09-15T16:48:49.960Z|22763|binding|INFO|9a6cc189-0934-4468-97ae-09f90fa4598d:
>>>>>>>> Claiming 56:6f:77:61:00:06
>>>>>>>> 2020-09-15T16:48:49.960Z|22764|binding|INFO|Claiming
lport
>>>>>>>> 16162721-c815-4cd8-ab57-f22e6e482c7f for this chassis.
>>>>>>>>
2020-09-15T16:48:49.960Z|22765|binding|INFO|16162721-c815-4cd8-ab57-f22e6e482c7f:
>>>>>>>> Claiming 56:6f:77:61:00:03
>>>>>>>> 2020-09-15T16:48:49.960Z|22766|binding|INFO|Claiming
lport
>>>>>>>> b88de6e4-6d77-4e42-b734-4cc676728910 for this chassis.
>>>>>>>>
2020-09-15T16:48:49.960Z|22767|binding|INFO|b88de6e4-6d77-4e42-b734-4cc676728910:
>>>>>>>> Claiming 56:6f:77:61:00:15
>>>>>>>> 2020-09-15T16:48:49.960Z|22768|binding|INFO|Claiming
lport
>>>>>>>> b7ff5f2b-4bb4-4250-8ad8-8a7e19d2b4c7 for this chassis.
>>>>>>>>
2020-09-15T16:48:49.960Z|22769|binding|INFO|b7ff5f2b-4bb4-4250-8ad8-8a7e19d2b4c7:
>>>>>>>> Claiming 56:6f:77:61:00:0d
>>>>>>>> 2020-09-15T16:48:49.960Z|22770|binding|INFO|Claiming
lport
>>>>>>>> 5d03a7a5-82a1-40f9-b50c-353a26167fa3 for this chassis.
>>>>>>>>
2020-09-15T16:48:49.960Z|22771|binding|INFO|5d03a7a5-82a1-40f9-b50c-353a26167fa3:
>>>>>>>> Claiming 56:6f:77:61:00:02
>>>>>>>> 2020-09-15T16:48:49.960Z|22772|binding|INFO|Claiming
lport
>>>>>>>> 12d829c3-64eb-44bc-a0bd-d7219991f35f for this chassis.
>>>>>>>>
2020-09-15T16:48:49.960Z|22773|binding|INFO|12d829c3-64eb-44bc-a0bd-d7219991f35f:
>>>>>>>> Claiming 56:6f:77:61:00:1c
>>>>>>>>
>>>>>>>>
>>>>>>>> And this repeats forever.
>>>>>>>>
>>>>>>>>
>>>>>>> Looks like the southbound db is confused.
>>>>>>>
>>>>>>> Can you try to delete all chassis listed by
>>>>>>> sudo ovn-sbctl show
>>>>>>> via
>>>>>>> sudo /usr/share/ovirt-provider-ovn/scripts/remove_chassis.sh
>>>>>>> dev-host0
>>>>>>> ?
>>>>>>> if the script remove_chassis.sh is not installed, you can
use
>>>>>>>
>>>>>>>
https://github.com/oVirt/ovirt-provider-ovn/blob/master/provider/scripts/...
>>>>>>> instead.
>>>>>>>
>>>>>>> Can you please also share the output of
>>>>>>> ovs-vsctl list Interface
>>>>>>> on the host which produced the logfile above?
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>> The connections to ovn-sbctl is ok and the geneve tunnels
are
>>>>>>>> depicted under ovs-vsctl ok.
>>>>>>>> VMs still not able to ping each other.
>>>>>>>>
>>>>>>>> On Tue, Sep 15, 2020 at 7:22 PM Dominik Holler
<dholler(a)redhat.com>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On Tue, Sep 15, 2020 at 6:18 PM Konstantinos Betsis
<
>>>>>>>>> k.betsis(a)gmail.com> wrote:
>>>>>>>>>
>>>>>>>>>> Hi Dominik
>>>>>>>>>>
>>>>>>>>>> Fixed the issue.
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>> Thanks.
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>> I believe the
/etc/ovirt-provider-ovn/conf.d/10-setup-ovirt-provider-ovn.conf
>>>>>>>>>> needed update also.
>>>>>>>>>> The package is upgraded to the latest version.
>>>>>>>>>>
>>>>>>>>>> Once the provider was updated with the following
it functioned
>>>>>>>>>> perfectly:
>>>>>>>>>>
>>>>>>>>>> Name: ovirt-provider-ovn
>>>>>>>>>> Description: oVirt network provider for OVN
>>>>>>>>>> Type: External Network Provider
>>>>>>>>>> Network Plugin: oVirt Network Provider for OVN
>>>>>>>>>> Automatic Synchronization: Checked
>>>>>>>>>> Unmanaged: Unchecked
>>>>>>>>>> Provider URL:
https:dc02-ovirt01.testdomain.com:9696
>>>>>>>>>> Requires Authentication: Checked
>>>>>>>>>> Username: admin@internal
>>>>>>>>>> Password: "The admin password"
>>>>>>>>>> Protocol: HTTPS
>>>>>>>>>> Host Name:
dc02-ovirt01.testdomain.com
>>>>>>>>>> API Port: 35357
>>>>>>>>>> API Version: v2.0
>>>>>>>>>> Tenant Name: "Empty"
>>>>>>>>>>
>>>>>>>>>> For some reason the TLS certificate was in
conflict with the ovn
>>>>>>>>>> provider details, i would bet the
"host" entry.
>>>>>>>>>>
>>>>>>>>>> So now geneve tunnels are established.
>>>>>>>>>> OVN provider is working.
>>>>>>>>>>
>>>>>>>>>> But VMs still do not communicated on the same VM
network
>>>>>>>>>> spanning different hosts.
>>>>>>>>>>
>>>>>>>>>> So if we have a VM network test-net on both
dc01-host01 and
>>>>>>>>>> dc01-host02 and each host has a VM with IP
addresses on the same network,
>>>>>>>>>> VMs on the same VM network should communicate
directly.
>>>>>>>>>> But traffic does not reach each other.
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>> Can you create a new external network, with port
security
>>>>>>>>> disabled, and an IPv4 subnet?
>>>>>>>>> If the VMs get an IP address via DHCP, ovn is
working, and should
>>>>>>>>> be able to ping each other, too.
>>>>>>>>> If not, there should be a helpful entry in the
ovn-controller.log
>>>>>>>>> of the host the VM is running.
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>> On Tue, Sep 15, 2020 at 7:07 PM Dominik Holler
<
>>>>>>>>>> dholler(a)redhat.com> wrote:
>>>>>>>>>>
>>>>>>>>>>> Can you try again with:
>>>>>>>>>>>
>>>>>>>>>>> [OVN REMOTE]
>>>>>>>>>>> ovn-remote=ssl:127.0.0.1:6641
>>>>>>>>>>> [SSL]
>>>>>>>>>>> https-enabled=false
>>>>>>>>>>> ssl-cacert-file=/etc/pki/ovirt-engine/ca.pem
>>>>>>>>>>>
ssl-cert-file=/etc/pki/ovirt-engine/certs/ovirt-provider-ovn.cer
>>>>>>>>>>>
>>>>>>>>>>>
ssl-key-file=/etc/pki/ovirt-engine/keys/ovirt-provider-ovn.key.nopass
>>>>>>>>>>> [OVIRT]
>>>>>>>>>>> ovirt-sso-client-secret=*random_test*
>>>>>>>>>>>
ovirt-host=https://dc02-ovirt01.testdomain.com:443
>>>>>>>>>>> <
https://dc02-ovirt01.testdomain.com/>
>>>>>>>>>>> ovirt-sso-client-id=ovirt-provider-ovn
>>>>>>>>>>>
ovirt-ca-file=/etc/pki/ovirt-engine/apache-ca.pem
>>>>>>>>>>> [NETWORK]
>>>>>>>>>>> port-security-enabled-default=True
>>>>>>>>>>> [PROVIDER]
>>>>>>>>>>>
>>>>>>>>>>>
provider-host=dc02-ovirt01.testdomain.com
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> Please note that the should match the HTTP or
HTTPS in the of
>>>>>>>>>>> the ovirt-prover-ovn configuration in oVirt
Engine.
>>>>>>>>>>> So if the ovirt-provider-ovn entity in Engine
is on HTTP, the
>>>>>>>>>>> config file should use
>>>>>>>>>>> https-enabled=false
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> On Tue, Sep 15, 2020 at 5:56 PM Konstantinos
Betsis <
>>>>>>>>>>> k.betsis(a)gmail.com> wrote:
>>>>>>>>>>>
>>>>>>>>>>>> This is the updated one:
>>>>>>>>>>>>
>>>>>>>>>>>> # This file is automatically generated by
engine-setup. Please
>>>>>>>>>>>> do not edit manually
>>>>>>>>>>>> [OVN REMOTE]
>>>>>>>>>>>> ovn-remote=ssl:127.0.0.1:6641
>>>>>>>>>>>> [SSL]
>>>>>>>>>>>> https-enabled=true
>>>>>>>>>>>>
ssl-cacert-file=/etc/pki/ovirt-engine/ca.pem
>>>>>>>>>>>>
>>>>>>>>>>>>
ssl-cert-file=/etc/pki/ovirt-engine/certs/ovirt-provider-ovn.cer
>>>>>>>>>>>>
>>>>>>>>>>>>
ssl-key-file=/etc/pki/ovirt-engine/keys/ovirt-provider-ovn.key.nopass
>>>>>>>>>>>> [OVIRT]
>>>>>>>>>>>> ovirt-sso-client-secret=*random_text*
>>>>>>>>>>>>
ovirt-host=https://dc02-ovirt01.testdomain.com:443
>>>>>>>>>>>> ovirt-sso-client-id=ovirt-provider-ovn
>>>>>>>>>>>>
ovirt-ca-file=/etc/pki/ovirt-engine/apache-ca.pem
>>>>>>>>>>>> [NETWORK]
>>>>>>>>>>>> port-security-enabled-default=True
>>>>>>>>>>>> [PROVIDER]
>>>>>>>>>>>>
provider-host=dc02-ovirt01.testdomain.com
>>>>>>>>>>>> [AUTH]
>>>>>>>>>>>>
auth-plugin=auth.plugins.static_token:NoAuthPlugin
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> However, it still does not connect.
>>>>>>>>>>>> It prompts for the certificate but then
fails and prompts to
>>>>>>>>>>>> see the log but the
ovirt-provider-ovn.log does not list anything.
>>>>>>>>>>>>
>>>>>>>>>>>> Yes we've got ovirt for about a year
now from about version 4.1
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>> This might explain the trouble. Upgrade of
ovirt-provider-ovn
>>>>>>>>>>> should work flawlessly starting from oVirt
4.2.
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>> On Tue, Sep 15, 2020 at 6:44 PM Dominik
Holler <
>>>>>>>>>>>> dholler(a)redhat.com> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> On Tue, Sep 15, 2020 at 5:34 PM
Konstantinos Betsis <
>>>>>>>>>>>>> k.betsis(a)gmail.com> wrote:
>>>>>>>>>>>>>
>>>>>>>>>>>>>> There is a file with the below
entries
>>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> Impressive, do you know when this
config file was created and
>>>>>>>>>>>>> if it was manually modified?
>>>>>>>>>>>>> Is this an upgrade from oVirt 4.1?
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>> [root@dc02-ovirt01 log]# cat
>>>>>>>>>>>>>>
/etc/ovirt-provider-ovn/conf.d/10-setup-ovirt-provider-ovn.conf
>>>>>>>>>>>>>> # This file is automatically
generated by engine-setup.
>>>>>>>>>>>>>> Please do not edit manually
>>>>>>>>>>>>>> [OVN REMOTE]
>>>>>>>>>>>>>> ovn-remote=tcp:127.0.0.1:6641
>>>>>>>>>>>>>> [SSL]
>>>>>>>>>>>>>> https-enabled=false
>>>>>>>>>>>>>>
ssl-cacert-file=/etc/pki/ovirt-engine/ca.pem
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
ssl-cert-file=/etc/pki/ovirt-engine/certs/ovirt-provider-ovn.cer
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
ssl-key-file=/etc/pki/ovirt-engine/keys/ovirt-provider-ovn.key.nopass
>>>>>>>>>>>>>> [OVIRT]
>>>>>>>>>>>>>>
ovirt-sso-client-secret=*random_test*
>>>>>>>>>>>>>>
ovirt-host=https://dc02-ovirt01.testdomain.com:443
>>>>>>>>>>>>>>
ovirt-sso-client-id=ovirt-provider-ovn
>>>>>>>>>>>>>>
ovirt-ca-file=/etc/pki/ovirt-engine/apache-ca.pem
>>>>>>>>>>>>>> [NETWORK]
>>>>>>>>>>>>>>
port-security-enabled-default=True
>>>>>>>>>>>>>> [PROVIDER]
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
provider-host=dc02-ovirt01.testdomain.com
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> The only entry missing is the
[AUTH] and under [SSL] the
>>>>>>>>>>>>>> https-enabled is false. Should I
edit this in this file or is this going to
>>>>>>>>>>>>>> break everything?
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>> Changing the file should improve, but
better create a backup
>>>>>>>>>>>>> into another diretory before
modification.
>>>>>>>>>>>>> The only required change is
>>>>>>>>>>>>> from
>>>>>>>>>>>>> ovn-remote=tcp:127.0.0.1:6641
>>>>>>>>>>>>> to
>>>>>>>>>>>>> ovn-remote=ssl:127.0.0.1:6641
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>> On Tue, Sep 15, 2020 at 6:27 PM
Dominik Holler <
>>>>>>>>>>>>>> dholler(a)redhat.com> wrote:
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> On Tue, Sep 15, 2020 at 5:11
PM Konstantinos Betsis <
>>>>>>>>>>>>>>> k.betsis(a)gmail.com>
wrote:
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Hi Dominik
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> That immediately fixed
the geneve tunnels between all
>>>>>>>>>>>>>>>> hosts.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> thanks for the feedback.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> However, the ovn provider
is not broken.
>>>>>>>>>>>>>>>> After fixing the networks
we tried to move a VM to the
>>>>>>>>>>>>>>>> DC01-host01 so we powered
it down and simply configured it to run on
>>>>>>>>>>>>>>>> dc01-node01.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> While checking the logs
on the ovirt engine i noticed the
>>>>>>>>>>>>>>>> below:
>>>>>>>>>>>>>>>> Failed to synchronize
networks of Provider
>>>>>>>>>>>>>>>> ovirt-provider-ovn.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> The ovn-provider
configure on the engine is the below:
>>>>>>>>>>>>>>>> Name: ovirt-provider-ovn
>>>>>>>>>>>>>>>> Description: oVirt
network provider for OVN
>>>>>>>>>>>>>>>> Type: External Network
Provider
>>>>>>>>>>>>>>>> Network Plugin: oVirt
Network Provider for OVN
>>>>>>>>>>>>>>>> Automatic
Synchronization: Checked
>>>>>>>>>>>>>>>> Unmanaged: Unchecked
>>>>>>>>>>>>>>>> Provider URL:
http:localhost:9696
>>>>>>>>>>>>>>>> Requires Authentication:
Checked
>>>>>>>>>>>>>>>> Username: admin@internal
>>>>>>>>>>>>>>>> Password: "The admin
password"
>>>>>>>>>>>>>>>> Protocol: hTTP
>>>>>>>>>>>>>>>> Host Name: dc02-ovirt01
>>>>>>>>>>>>>>>> API Port: 35357
>>>>>>>>>>>>>>>> API Version: v2.0
>>>>>>>>>>>>>>>> Tenant Name:
"Empty"
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> In the past this was
deleted by an engineer and recreated
>>>>>>>>>>>>>>>> as per the documentation,
and it worked. Do we need to update something due
>>>>>>>>>>>>>>>> to the SSL on the ovn?
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Is there a file in
/etc/ovirt-provider-ovn/conf.d/ ?
>>>>>>>>>>>>>>> engine-setup should have
created one.
>>>>>>>>>>>>>>> If the file is missing, for
testing purposes, you can
>>>>>>>>>>>>>>> create a
>>>>>>>>>>>>>>> file
/etc/ovirt-provider-ovn/conf.d/00-setup-ovirt-provider-ovn-test.conf :
>>>>>>>>>>>>>>> [PROVIDER]
>>>>>>>>>>>>>>>
provider-host=REPLACE_WITH_FQDN
>>>>>>>>>>>>>>> [SSL]
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
ssl-cert-file=/etc/pki/ovirt-engine/certs/ovirt-provider-ovn.cer
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
ssl-key-file=/etc/pki/ovirt-engine/keys/ovirt-provider-ovn.key.nopass
>>>>>>>>>>>>>>>
ssl-cacert-file=/etc/pki/ovirt-engine/ca.pem
>>>>>>>>>>>>>>> https-enabled=true
>>>>>>>>>>>>>>> [OVN REMOTE]
>>>>>>>>>>>>>>>
ovn-remote=ssl:127.0.0.1:6641
>>>>>>>>>>>>>>> [AUTH]
>>>>>>>>>>>>>>>
auth-plugin=auth.plugins.static_token:NoAuthPlugin
>>>>>>>>>>>>>>> [NETWORK]
>>>>>>>>>>>>>>>
port-security-enabled-default=True
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> and restart the
ovirt-provider-ovn service.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> From the ovn-provider
logs the below is generated after a
>>>>>>>>>>>>>>>> service restart and when
the start VM is triggered
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> 2020-09-15 15:07:33,579
root Starting server
>>>>>>>>>>>>>>>> 2020-09-15 15:07:33,579
root Version: 1.2.29-1
>>>>>>>>>>>>>>>> 2020-09-15 15:07:33,579
root Build date: 20191217125241
>>>>>>>>>>>>>>>> 2020-09-15 15:07:33,579
root Githash: cb5a80d
>>>>>>>>>>>>>>>> 2020-09-15 15:08:26,582
root From: ::ffff:127.0.0.1:59980
>>>>>>>>>>>>>>>> Request: GET /v2.0/ports
>>>>>>>>>>>>>>>> 2020-09-15 15:08:26,582
root Could not retrieve schema
>>>>>>>>>>>>>>>> from tcp:127.0.0.1:6641:
Unknown error -1
>>>>>>>>>>>>>>>> Traceback (most recent
call last):
>>>>>>>>>>>>>>>> File
>>>>>>>>>>>>>>>>
"/usr/share/ovirt-provider-ovn/handlers/base_handler.py", line 138, in
>>>>>>>>>>>>>>>> _handle_request
>>>>>>>>>>>>>>>> method, path_parts,
content
>>>>>>>>>>>>>>>> File
>>>>>>>>>>>>>>>>
"/usr/share/ovirt-provider-ovn/handlers/selecting_handler.py", line 175, in
>>>>>>>>>>>>>>>> handle_request
>>>>>>>>>>>>>>>> return
self.call_response_handler(handler, content,
>>>>>>>>>>>>>>>> parameters)
>>>>>>>>>>>>>>>> File
>>>>>>>>>>>>>>>>
"/usr/share/ovirt-provider-ovn/handlers/neutron.py", line 35, in
>>>>>>>>>>>>>>>> call_response_handler
>>>>>>>>>>>>>>>> with NeutronApi() as
ovn_north:
>>>>>>>>>>>>>>>> File
>>>>>>>>>>>>>>>>
"/usr/share/ovirt-provider-ovn/neutron/neutron_api.py", line 95, in __init__
>>>>>>>>>>>>>>>> self.ovsidl, self.idl
= ovn_connection.connect()
>>>>>>>>>>>>>>>> File
"/usr/share/ovirt-provider-ovn/ovn_connection.py",
>>>>>>>>>>>>>>>> line 46, in connect
>>>>>>>>>>>>>>>>
ovnconst.OVN_NORTHBOUND
>>>>>>>>>>>>>>>> File
>>>>>>>>>>>>>>>>
"/usr/lib/python2.7/site-packages/ovsdbapp/backend/ovs_idl/connection.py",
>>>>>>>>>>>>>>>> line 127, in from_server
>>>>>>>>>>>>>>>> helper =
idlutils.get_schema_helper(connection_string,
>>>>>>>>>>>>>>>> schema_name)
>>>>>>>>>>>>>>>> File
>>>>>>>>>>>>>>>>
"/usr/lib/python2.7/site-packages/ovsdbapp/backend/ovs_idl/idlutils.py",
>>>>>>>>>>>>>>>> line 128, in
get_schema_helper
>>>>>>>>>>>>>>>> 'err':
os.strerror(err)})
>>>>>>>>>>>>>>>> Exception: Could not
retrieve schema from tcp:
>>>>>>>>>>>>>>>> 127.0.0.1:6641: Unknown
error -1
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> When i update the ovn
provider from the GUI to have
>>>>>>>>>>>>>>>>
https://localhost:9696/
and HTTPS as the protocol the
>>>>>>>>>>>>>>>> test fails.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> On Tue, Sep 15, 2020 at
5:35 PM Dominik Holler <
>>>>>>>>>>>>>>>> dholler(a)redhat.com>
wrote:
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> On Mon, Sep 14, 2020
at 9:25 AM Konstantinos Betsis <
>>>>>>>>>>>>>>>>>
k.betsis(a)gmail.com> wrote:
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Hi Dominik
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> When these
commands are used on the ovirt-engine host
>>>>>>>>>>>>>>>>>> the output is the
one depicted in your email.
>>>>>>>>>>>>>>>>>> For your
reference see also below:
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
[root@ath01-ovirt01 certs]# ovn-nbctl get-ssl
>>>>>>>>>>>>>>>>>> Private key:
>>>>>>>>>>>>>>>>>>
/etc/pki/ovirt-engine/keys/ovn-ndb.key.nopass
>>>>>>>>>>>>>>>>>> Certificate:
/etc/pki/ovirt-engine/certs/ovn-ndb.cer
>>>>>>>>>>>>>>>>>> CA Certificate:
/etc/pki/ovirt-engine/ca.pem
>>>>>>>>>>>>>>>>>> Bootstrap: false
>>>>>>>>>>>>>>>>>>
[root@ath01-ovirt01 certs]# ovn-nbctl get-connection
>>>>>>>>>>>>>>>>>> ptcp:6641
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
[root@ath01-ovirt01 certs]# ovn-sbctl get-ssl
>>>>>>>>>>>>>>>>>> Private key:
>>>>>>>>>>>>>>>>>>
/etc/pki/ovirt-engine/keys/ovn-sdb.key.nopass
>>>>>>>>>>>>>>>>>> Certificate:
/etc/pki/ovirt-engine/certs/ovn-sdb.cer
>>>>>>>>>>>>>>>>>> CA Certificate:
/etc/pki/ovirt-engine/ca.pem
>>>>>>>>>>>>>>>>>> Bootstrap: false
>>>>>>>>>>>>>>>>>>
[root@ath01-ovirt01 certs]# ovn-sbctl get-connection
>>>>>>>>>>>>>>>>>> read-write
role="" ptcp:6642
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> ^^^ the line above
points to the problem: ovn-central is
>>>>>>>>>>>>>>>>> configured to use
plain TCP without ssl.
>>>>>>>>>>>>>>>>> engine-setup usually
configures ovn-central to use SSL.
>>>>>>>>>>>>>>>>> That the files
/etc/pki/ovirt-engine/keys/ovn-* exist, shows,
>>>>>>>>>>>>>>>>> that engine-setup was
triggered correctly. Looks like the
>>>>>>>>>>>>>>>>> ovn db was dropped
somehow, this should not happen.
>>>>>>>>>>>>>>>>> This can be fixed
manually by executing the following
>>>>>>>>>>>>>>>>> commands on
engine's machine:
>>>>>>>>>>>>>>>>> ovn-nbctl set-ssl
>>>>>>>>>>>>>>>>>
/etc/pki/ovirt-engine/keys/ovn-ndb.key.nopass
>>>>>>>>>>>>>>>>>
/etc/pki/ovirt-engine/certs/ovn-ndb.cer /etc/pki/ovirt-engine/ca.pem
>>>>>>>>>>>>>>>>> ovn-nbctl
set-connection pssl:6641
>>>>>>>>>>>>>>>>> ovn-sbctl set-ssl
>>>>>>>>>>>>>>>>>
/etc/pki/ovirt-engine/keys/ovn-sdb.key.nopass
>>>>>>>>>>>>>>>>>
/etc/pki/ovirt-engine/certs/ovn-sdb.cer /etc/pki/ovirt-engine/ca.pem
>>>>>>>>>>>>>>>>> ovn-sbctl
set-connection pssl:6642
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> The
/var/log/openvswitch/ovn-controller.log on the hosts
>>>>>>>>>>>>>>>>> should tell that
br-int.mgmt is connected now.
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
[root@ath01-ovirt01 certs]# ls -l
>>>>>>>>>>>>>>>>>>
/etc/pki/ovirt-engine/keys/ovn-*
>>>>>>>>>>>>>>>>>> -rw-r-----. 1
root hugetlbfs 1828 Jun 25 11:08
>>>>>>>>>>>>>>>>>>
/etc/pki/ovirt-engine/keys/ovn-ndb.key.nopass
>>>>>>>>>>>>>>>>>> -rw-------. 1
root root 2893 Jun 25 11:08
>>>>>>>>>>>>>>>>>>
/etc/pki/ovirt-engine/keys/ovn-ndb.p12
>>>>>>>>>>>>>>>>>> -rw-r-----. 1
root hugetlbfs 1828 Jun 25 11:08
>>>>>>>>>>>>>>>>>>
/etc/pki/ovirt-engine/keys/ovn-sdb.key.nopass
>>>>>>>>>>>>>>>>>> -rw-------. 1
root root 2893 Jun 25 11:08
>>>>>>>>>>>>>>>>>>
/etc/pki/ovirt-engine/keys/ovn-sdb.p12
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> When i try the
above commands on the node hosts the
>>>>>>>>>>>>>>>>>> following
happens:
>>>>>>>>>>>>>>>>>> ovn-nbctl get-ssl
/ get-connection
>>>>>>>>>>>>>>>>>> ovn-nbctl:
unix:/var/run/openvswitch/ovnnb_db.sock:
>>>>>>>>>>>>>>>>>> database
connection failed (No such file or directory)
>>>>>>>>>>>>>>>>>> The above i
believe is expected since no northbound
>>>>>>>>>>>>>>>>>> connections
should be established from the host nodes.
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> ovn-sbctl get-ssl
/get-connection
>>>>>>>>>>>>>>>>>> The output is
stuck till i terminate it.
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Yes, the ovn-*
commands works only on engine's machine,
>>>>>>>>>>>>>>>>> which has the role
ovn-central.
>>>>>>>>>>>>>>>>> On the hosts, there
is only the ovn-controller, which
>>>>>>>>>>>>>>>>> connects the ovn
southbound to openvswitch on the host.
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> For the requested
logs the below are found in the
>>>>>>>>>>>>>>>>>>
ovsdb-server-sb.log
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
2020-09-14T07:18:38.187Z|219636|reconnect|WARN|tcp:DC02-host01:33146:
>>>>>>>>>>>>>>>>>> connection
dropped (Protocol error)
>>>>>>>>>>>>>>>>>>
2020-09-14T07:18:41.946Z|219637|reconnect|WARN|tcp:DC01-host01:51188:
>>>>>>>>>>>>>>>>>> connection
dropped (Protocol error)
>>>>>>>>>>>>>>>>>>
2020-09-14T07:18:43.033Z|219638|reconnect|WARN|tcp:DC01-host02:37044:
>>>>>>>>>>>>>>>>>> connection
dropped (Protocol error)
>>>>>>>>>>>>>>>>>>
2020-09-14T07:18:46.198Z|219639|reconnect|WARN|tcp:DC02-host01:33148:
>>>>>>>>>>>>>>>>>> connection
dropped (Protocol error)
>>>>>>>>>>>>>>>>>>
2020-09-14T07:18:50.069Z|219640|jsonrpc|WARN|Dropped 4
>>>>>>>>>>>>>>>>>> log messages in
last 12 seconds (most recently, 4 seconds ago) due to
>>>>>>>>>>>>>>>>>> excessive rate
>>>>>>>>>>>>>>>>>>
2020-09-14T07:18:50.069Z|219641|jsonrpc|WARN|tcp:DC01-host01:51190:
>>>>>>>>>>>>>>>>>> error parsing
stream: line 0, column 0, byte 0: invalid character U+0016
>>>>>>>>>>>>>>>>>>
2020-09-14T07:18:50.069Z|219642|jsonrpc|WARN|Dropped 4
>>>>>>>>>>>>>>>>>> log messages in
last 12 seconds (most recently, 4 seconds ago) due to
>>>>>>>>>>>>>>>>>> excessive rate
>>>>>>>>>>>>>>>>>>
2020-09-14T07:18:50.069Z|219643|jsonrpc|WARN|tcp:DC01-host01:51190:
>>>>>>>>>>>>>>>>>> received SSL data
on JSON-RPC channel
>>>>>>>>>>>>>>>>>>
2020-09-14T07:18:50.070Z|219644|reconnect|WARN|tcp:DC01-host01:51190:
>>>>>>>>>>>>>>>>>> connection
dropped (Protocol error)
>>>>>>>>>>>>>>>>>>
2020-09-14T07:18:51.147Z|219645|reconnect|WARN|tcp:DC01-host02:37046:
>>>>>>>>>>>>>>>>>> connection
dropped (Protocol error)
>>>>>>>>>>>>>>>>>>
2020-09-14T07:18:54.209Z|219646|reconnect|WARN|tcp:DC02-host01:33150:
>>>>>>>>>>>>>>>>>> connection
dropped (Protocol error)
>>>>>>>>>>>>>>>>>>
2020-09-14T07:18:58.192Z|219647|reconnect|WARN|tcp:DC01-host01:51192:
>>>>>>>>>>>>>>>>>> connection
dropped (Protocol error)
>>>>>>>>>>>>>>>>>>
2020-09-14T07:18:59.262Z|219648|jsonrpc|WARN|Dropped 3
>>>>>>>>>>>>>>>>>> log messages in
last 8 seconds (most recently, 1 seconds ago) due to
>>>>>>>>>>>>>>>>>> excessive rate
>>>>>>>>>>>>>>>>>>
2020-09-14T07:18:59.262Z|219649|jsonrpc|WARN|tcp:DC01-host02:37048:
>>>>>>>>>>>>>>>>>> error parsing
stream: line 0, column 0, byte 0: invalid character U+0016
>>>>>>>>>>>>>>>>>>
2020-09-14T07:18:59.263Z|219650|jsonrpc|WARN|Dropped 3
>>>>>>>>>>>>>>>>>> log messages in
last 8 seconds (most recently, 1 seconds ago) due to
>>>>>>>>>>>>>>>>>> excessive rate
>>>>>>>>>>>>>>>>>>
2020-09-14T07:18:59.263Z|219651|jsonrpc|WARN|tcp:DC01-host02:37048:
>>>>>>>>>>>>>>>>>> received SSL data
on JSON-RPC channel
>>>>>>>>>>>>>>>>>>
2020-09-14T07:18:59.263Z|219652|reconnect|WARN|tcp:DC01-host02:37048:
>>>>>>>>>>>>>>>>>> connection
dropped (Protocol error)
>>>>>>>>>>>>>>>>>>
2020-09-14T07:19:02.220Z|219653|reconnect|WARN|tcp:DC02-host01:33152:
>>>>>>>>>>>>>>>>>> connection
dropped (Protocol error)
>>>>>>>>>>>>>>>>>>
2020-09-14T07:19:06.316Z|219654|reconnect|WARN|tcp:DC01-host01:51194:
>>>>>>>>>>>>>>>>>> connection
dropped (Protocol error)
>>>>>>>>>>>>>>>>>>
2020-09-14T07:19:07.386Z|219655|reconnect|WARN|tcp:DC01-host02:37050:
>>>>>>>>>>>>>>>>>> connection
dropped (Protocol error)
>>>>>>>>>>>>>>>>>>
2020-09-14T07:19:10.232Z|219656|reconnect|WARN|tcp:DC02-host01:33154:
>>>>>>>>>>>>>>>>>> connection
dropped (Protocol error)
>>>>>>>>>>>>>>>>>>
2020-09-14T07:19:14.439Z|219657|jsonrpc|WARN|Dropped 4
>>>>>>>>>>>>>>>>>> log messages in
last 12 seconds (most recently, 4 seconds ago) due to
>>>>>>>>>>>>>>>>>> excessive rate
>>>>>>>>>>>>>>>>>>
2020-09-14T07:19:14.439Z|219658|jsonrpc|WARN|tcp:DC01-host01:51196:
>>>>>>>>>>>>>>>>>> error parsing
stream: line 0, column 0, byte 0: invalid character U+0016
>>>>>>>>>>>>>>>>>>
2020-09-14T07:19:14.439Z|219659|jsonrpc|WARN|Dropped 4
>>>>>>>>>>>>>>>>>> log messages in
last 12 seconds (most recently, 4 seconds ago) due to
>>>>>>>>>>>>>>>>>> excessive rate
>>>>>>>>>>>>>>>>>>
2020-09-14T07:19:14.439Z|219660|jsonrpc|WARN|tcp:DC01-host01:51196:
>>>>>>>>>>>>>>>>>> received SSL data
on JSON-RPC channel
>>>>>>>>>>>>>>>>>>
2020-09-14T07:19:14.440Z|219661|reconnect|WARN|tcp:DC01-host01:51196:
>>>>>>>>>>>>>>>>>> connection
dropped (Protocol error)
>>>>>>>>>>>>>>>>>>
2020-09-14T07:19:15.505Z|219662|reconnect|WARN|tcp:DC01-host02:37052:
>>>>>>>>>>>>>>>>>> connection
dropped (Protocol error)
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> How can we fix
these SSL errors?
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> I addressed this
above.
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> I thought vdsm
did the certificate provisioning on the
>>>>>>>>>>>>>>>>>> host nodes as to
communicate to the engine host node.
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Yes, this seems to
work in your scenario, just the SSL
>>>>>>>>>>>>>>>>> configuration on the
ovn-central was lost.
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> On Fri, Sep 11,
2020 at 6:39 PM Dominik Holler <
>>>>>>>>>>>>>>>>>>
dholler(a)redhat.com> wrote:
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Looks still
like the ovn-controller on the host
>>>>>>>>>>>>>>>>>>> has problems
communicating with ovn-southbound.
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Are there any
hints in /var/log/openvswitch/*.log,
>>>>>>>>>>>>>>>>>>> especially in
/var/log/openvswitch/ovsdb-server-sb.log ?
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Can you
please check the output of
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> ovn-nbctl
get-ssl
>>>>>>>>>>>>>>>>>>> ovn-nbctl
get-connection
>>>>>>>>>>>>>>>>>>> ovn-sbctl
get-ssl
>>>>>>>>>>>>>>>>>>> ovn-sbctl
get-connection
>>>>>>>>>>>>>>>>>>> ls -l
/etc/pki/ovirt-engine/keys/ovn-*
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> it should be
similar to
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
[root@ovirt-43 ~]# ovn-nbctl get-ssl
>>>>>>>>>>>>>>>>>>> Private key:
>>>>>>>>>>>>>>>>>>>
/etc/pki/ovirt-engine/keys/ovn-ndb.key.nopass
>>>>>>>>>>>>>>>>>>> Certificate:
/etc/pki/ovirt-engine/certs/ovn-ndb.cer
>>>>>>>>>>>>>>>>>>> CA
Certificate: /etc/pki/ovirt-engine/ca.pem
>>>>>>>>>>>>>>>>>>> Bootstrap:
false
>>>>>>>>>>>>>>>>>>>
[root@ovirt-43 ~]# ovn-nbctl get-connection
>>>>>>>>>>>>>>>>>>>
pssl:6641:[::]
>>>>>>>>>>>>>>>>>>>
[root@ovirt-43 ~]# ovn-sbctl get-ssl
>>>>>>>>>>>>>>>>>>> Private key:
>>>>>>>>>>>>>>>>>>>
/etc/pki/ovirt-engine/keys/ovn-sdb.key.nopass
>>>>>>>>>>>>>>>>>>> Certificate:
/etc/pki/ovirt-engine/certs/ovn-sdb.cer
>>>>>>>>>>>>>>>>>>> CA
Certificate: /etc/pki/ovirt-engine/ca.pem
>>>>>>>>>>>>>>>>>>> Bootstrap:
false
>>>>>>>>>>>>>>>>>>>
[root@ovirt-43 ~]# ovn-sbctl get-connection
>>>>>>>>>>>>>>>>>>> read-write
role="" pssl:6642:[::]
>>>>>>>>>>>>>>>>>>>
[root@ovirt-43 ~]# ls -l
>>>>>>>>>>>>>>>>>>>
/etc/pki/ovirt-engine/keys/ovn-*
>>>>>>>>>>>>>>>>>>> -rw-r-----. 1
root hugetlbfs 1828 Oct 14 2019
>>>>>>>>>>>>>>>>>>>
/etc/pki/ovirt-engine/keys/ovn-ndb.key.nopass
>>>>>>>>>>>>>>>>>>> -rw-------. 1
root root 2709 Oct 14 2019
>>>>>>>>>>>>>>>>>>>
/etc/pki/ovirt-engine/keys/ovn-ndb.p12
>>>>>>>>>>>>>>>>>>> -rw-r-----. 1
root hugetlbfs 1828 Oct 14 2019
>>>>>>>>>>>>>>>>>>>
/etc/pki/ovirt-engine/keys/ovn-sdb.key.nopass
>>>>>>>>>>>>>>>>>>> -rw-------. 1
root root 2709 Oct 14 2019
>>>>>>>>>>>>>>>>>>>
/etc/pki/ovirt-engine/keys/ovn-sdb.p12
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> On Fri, Sep
11, 2020 at 1:10 PM Konstantinos Betsis <
>>>>>>>>>>>>>>>>>>>
k.betsis(a)gmail.com> wrote:
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> I did a
restart of the ovn-controller, this is the
>>>>>>>>>>>>>>>>>>>> output of
the ovn-controller.log
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>
2020-09-11T10:54:07.566Z|00001|vlog|INFO|opened log
>>>>>>>>>>>>>>>>>>>> file
/var/log/openvswitch/ovn-controller.log
>>>>>>>>>>>>>>>>>>>>
2020-09-11T10:54:07.568Z|00002|reconnect|INFO|unix:/var/run/openvswitch/db.sock:
>>>>>>>>>>>>>>>>>>>>
connecting...
>>>>>>>>>>>>>>>>>>>>
2020-09-11T10:54:07.568Z|00003|reconnect|INFO|unix:/var/run/openvswitch/db.sock:
>>>>>>>>>>>>>>>>>>>>
connected
>>>>>>>>>>>>>>>>>>>>
2020-09-11T10:54:07.570Z|00004|main|INFO|OVS IDL
>>>>>>>>>>>>>>>>>>>>
reconnected, force recompute.
>>>>>>>>>>>>>>>>>>>>
2020-09-11T10:54:07.571Z|00005|reconnect|INFO|ssl:OVIRT_ENGINE_IP:6642:
>>>>>>>>>>>>>>>>>>>>
connecting...
>>>>>>>>>>>>>>>>>>>>
2020-09-11T10:54:07.571Z|00006|main|INFO|OVNSB IDL
>>>>>>>>>>>>>>>>>>>>
reconnected, force recompute.
>>>>>>>>>>>>>>>>>>>>
2020-09-11T10:54:07.685Z|00007|stream_ssl|WARN|SSL_connect:
>>>>>>>>>>>>>>>>>>>>
unexpected SSL connection close
>>>>>>>>>>>>>>>>>>>>
2020-09-11T10:54:07.685Z|00008|reconnect|INFO|ssl:OVIRT_ENGINE_IP:6642:
>>>>>>>>>>>>>>>>>>>>
connection attempt failed (Protocol error)
>>>>>>>>>>>>>>>>>>>>
2020-09-11T10:54:08.685Z|00009|reconnect|INFO|ssl:OVIRT_ENGINE_IP:6642:
>>>>>>>>>>>>>>>>>>>>
connecting...
>>>>>>>>>>>>>>>>>>>>
2020-09-11T10:54:08.800Z|00010|stream_ssl|WARN|SSL_connect:
>>>>>>>>>>>>>>>>>>>>
unexpected SSL connection close
>>>>>>>>>>>>>>>>>>>>
2020-09-11T10:54:08.800Z|00011|reconnect|INFO|ssl:OVIRT_ENGINE_IP:6642:
>>>>>>>>>>>>>>>>>>>>
connection attempt failed (Protocol error)
>>>>>>>>>>>>>>>>>>>>
2020-09-11T10:54:08.800Z|00012|reconnect|INFO|ssl:OVIRT_ENGINE_IP:6642:
>>>>>>>>>>>>>>>>>>>> waiting 2
seconds before reconnect
>>>>>>>>>>>>>>>>>>>>
2020-09-11T10:54:10.802Z|00013|reconnect|INFO|ssl:OVIRT_ENGINE_IP:6642:
>>>>>>>>>>>>>>>>>>>>
connecting...
>>>>>>>>>>>>>>>>>>>>
2020-09-11T10:54:10.917Z|00014|stream_ssl|WARN|SSL_connect:
>>>>>>>>>>>>>>>>>>>>
unexpected SSL connection close
>>>>>>>>>>>>>>>>>>>>
2020-09-11T10:54:10.917Z|00015|reconnect|INFO|ssl:OVIRT_ENGINE_IP:6642:
>>>>>>>>>>>>>>>>>>>>
connection attempt failed (Protocol error)
>>>>>>>>>>>>>>>>>>>>
2020-09-11T10:54:10.917Z|00016|reconnect|INFO|ssl:OVIRT_ENGINE_IP:6642:
>>>>>>>>>>>>>>>>>>>> waiting 4
seconds before reconnect
>>>>>>>>>>>>>>>>>>>>
2020-09-11T10:54:14.921Z|00017|reconnect|INFO|ssl:OVIRT_ENGINE_IP:6642:
>>>>>>>>>>>>>>>>>>>>
connecting...
>>>>>>>>>>>>>>>>>>>>
2020-09-11T10:54:15.036Z|00018|stream_ssl|WARN|SSL_connect:
>>>>>>>>>>>>>>>>>>>>
unexpected SSL connection close
>>>>>>>>>>>>>>>>>>>>
2020-09-11T10:54:15.036Z|00019|reconnect|INFO|ssl:OVIRT_ENGINE_IP:6642:
>>>>>>>>>>>>>>>>>>>>
connection attempt failed (Protocol error)
>>>>>>>>>>>>>>>>>>>>
2020-09-11T10:54:15.036Z|00020|reconnect|INFO|ssl:OVIRT_ENGINE_IP:6642:
>>>>>>>>>>>>>>>>>>>>
continuing to reconnect in the background but suppressing further logging
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> I have
also done the vdsm-tool ovn-config
>>>>>>>>>>>>>>>>>>>>
OVIRT_ENGINE_IP OVIRTMGMT_NETWORK_DC
>>>>>>>>>>>>>>>>>>>> This is
how the OVIRT_ENGINE_IP is provided in the ovn
>>>>>>>>>>>>>>>>>>>>
controller, i can redo it if you wan.
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> After the
restart of the ovn-controller the OVIRT
>>>>>>>>>>>>>>>>>>>> ENGINE
still shows only two geneve connections one with DC01-host02 and
>>>>>>>>>>>>>>>>>>>>
DC02-host01.
>>>>>>>>>>>>>>>>>>>> Chassis
"c4b23834-aec7-4bf8-8be7-aa94a50a6144"
>>>>>>>>>>>>>>>>>>>>
hostname: "dc02-host01"
>>>>>>>>>>>>>>>>>>>> Encap
geneve
>>>>>>>>>>>>>>>>>>>>
ip: "DC02-host01_IP"
>>>>>>>>>>>>>>>>>>>>
options: {csum="true"}
>>>>>>>>>>>>>>>>>>>> Chassis
"be3abcc9-7358-4040-a37b-8d8a782f239c"
>>>>>>>>>>>>>>>>>>>>
hostname: "DC01-host02"
>>>>>>>>>>>>>>>>>>>> Encap
geneve
>>>>>>>>>>>>>>>>>>>>
ip: "DC01-host02"
>>>>>>>>>>>>>>>>>>>>
options: {csum="true"}
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> I've
re-done the vdsm-tool command and nothing
>>>>>>>>>>>>>>>>>>>>
changed.... again....with the same errors as the systemctl restart
>>>>>>>>>>>>>>>>>>>>
ovn-controller
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> On Fri,
Sep 11, 2020 at 1:49 PM Dominik Holler <
>>>>>>>>>>>>>>>>>>>>
dholler(a)redhat.com> wrote:
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>
Please include ovirt-users list in your reply, to
>>>>>>>>>>>>>>>>>>>>> share
the knowledge and experience with the community!
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> On
Fri, Sep 11, 2020 at 12:12 PM Konstantinos Betsis <
>>>>>>>>>>>>>>>>>>>>>
k.betsis(a)gmail.com> wrote:
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>
Ok below the output per node and DC
>>>>>>>>>>>>>>>>>>>>>>
DC01
>>>>>>>>>>>>>>>>>>>>>>
node01
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>
[root@dc01-node01 ~]# ovs-vsctl --no-wait get open
>>>>>>>>>>>>>>>>>>>>>> .
external-ids:ovn-remote
>>>>>>>>>>>>>>>>>>>>>>
"ssl:*OVIRT_ENGINE_IP*:6642"
>>>>>>>>>>>>>>>>>>>>>>
[root@ dc01-node01 ~]# ovs-vsctl --no-wait get open
>>>>>>>>>>>>>>>>>>>>>> .
external-ids:ovn-encap-type
>>>>>>>>>>>>>>>>>>>>>>
geneve
>>>>>>>>>>>>>>>>>>>>>>
[root@ dc01-node01 ~]# ovs-vsctl --no-wait get open
>>>>>>>>>>>>>>>>>>>>>> .
external-ids:ovn-encap-ip
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>
"*OVIRTMGMT_IP_DC01-NODE01*"
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>
node02
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>
[root@dc01-node02 ~]# ovs-vsctl --no-wait get open
>>>>>>>>>>>>>>>>>>>>>> .
external-ids:ovn-remote
>>>>>>>>>>>>>>>>>>>>>>
"ssl:*OVIRT_ENGINE_IP*:6642"
>>>>>>>>>>>>>>>>>>>>>>
[root@ dc01-node02 ~]# ovs-vsctl --no-wait get open
>>>>>>>>>>>>>>>>>>>>>> .
external-ids:ovn-encap-type
>>>>>>>>>>>>>>>>>>>>>>
geneve
>>>>>>>>>>>>>>>>>>>>>>
[root@ dc01-node02 ~]# ovs-vsctl --no-wait get open
>>>>>>>>>>>>>>>>>>>>>> .
external-ids:ovn-encap-ip
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>
"*OVIRTMGMT_IP_DC01-NODE02*"
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>
DC02
>>>>>>>>>>>>>>>>>>>>>>
node01
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>
[root@dc02-node01 ~]# ovs-vsctl --no-wait get open
>>>>>>>>>>>>>>>>>>>>>> .
external-ids:ovn-remote
>>>>>>>>>>>>>>>>>>>>>>
"ssl:*OVIRT_ENGINE_IP*:6642"
>>>>>>>>>>>>>>>>>>>>>>
[root@ dc02-node01 ~]# ovs-vsctl --no-wait get open
>>>>>>>>>>>>>>>>>>>>>> .
external-ids:ovn-encap-type
>>>>>>>>>>>>>>>>>>>>>>
geneve
>>>>>>>>>>>>>>>>>>>>>>
[root@ dc02-node01 ~]# ovs-vsctl --no-wait get open
>>>>>>>>>>>>>>>>>>>>>> .
external-ids:ovn-encap-ip
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>
"*OVIRTMGMT_IP_DC02-NODE01*"
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> Looks
good.
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>
DC01 node01 and node02 share the same VM networks
>>>>>>>>>>>>>>>>>>>>>>
and VMs deployed on top of them cannot talk to VM on the other hypervisor.
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> Maybe
there is a hint on ovn-controller.log on
>>>>>>>>>>>>>>>>>>>>>
dc01-node02 ? Maybe restarting ovn-controller creates more helpful log
>>>>>>>>>>>>>>>>>>>>>
messages?
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> You
can also try restart the ovn configuration on all
>>>>>>>>>>>>>>>>>>>>> hosts
by executing
>>>>>>>>>>>>>>>>>>>>>
vdsm-tool ovn-config OVIRT_ENGINE_IP
>>>>>>>>>>>>>>>>>>>>>
LOCAL_OVIRTMGMT_IP
>>>>>>>>>>>>>>>>>>>>> on
each host, this would trigger
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>
https://github.com/oVirt/ovirt-provider-ovn/blob/master/driver/scripts/se...
>>>>>>>>>>>>>>>>>>>>>
internally.
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>
So I would expect to see the same output for node01
>>>>>>>>>>>>>>>>>>>>>>
to have a geneve tunnel to node02 and vice versa.
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> Me
too.
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>
On Fri, Sep 11, 2020 at 12:14 PM Dominik Holler <
>>>>>>>>>>>>>>>>>>>>>>
dholler(a)redhat.com> wrote:
>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>
On Fri, Sep 11, 2020 at 10:53 AM Konstantinos
>>>>>>>>>>>>>>>>>>>>>>>
Betsis <k.betsis(a)gmail.com> wrote:
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>
Hi Dominik
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>
OVN is selected as the default network provider on
>>>>>>>>>>>>>>>>>>>>>>>>
the clusters and the hosts.
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>
sounds good.
>>>>>>>>>>>>>>>>>>>>>>>
This configuration is required already during the
>>>>>>>>>>>>>>>>>>>>>>>
host is added to oVirt Engine, because OVN is configured during this step.
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>
The "ovn-sbctl show" works on the ovirt engine and
>>>>>>>>>>>>>>>>>>>>>>>>
shows only two hosts, 1 per DC.
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>
Chassis "c4b23834-aec7-4bf8-8be7-aa94a50a6144"
>>>>>>>>>>>>>>>>>>>>>>>>
hostname: "dc01-node02"
>>>>>>>>>>>>>>>>>>>>>>>>
Encap geneve
>>>>>>>>>>>>>>>>>>>>>>>>
ip: "X.X.X.X"
>>>>>>>>>>>>>>>>>>>>>>>>
options: {csum="true"}
>>>>>>>>>>>>>>>>>>>>>>>>
Chassis "be3abcc9-7358-4040-a37b-8d8a782f239c"
>>>>>>>>>>>>>>>>>>>>>>>>
hostname: "dc02-node1"
>>>>>>>>>>>>>>>>>>>>>>>>
Encap geneve
>>>>>>>>>>>>>>>>>>>>>>>>
ip: "A.A.A.A"
>>>>>>>>>>>>>>>>>>>>>>>>
options: {csum="true"}
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>
The new node is not listed (dc01-node1).
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>
When executed on the nodes the same command
>>>>>>>>>>>>>>>>>>>>>>>>
(ovn-sbctl show) times-out on all nodes.....
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>
The output of the
>>>>>>>>>>>>>>>>>>>>>>>>
/var/log/openvswitch/ovn-conntroller.log lists on all logs
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>
2020-09-11T08:46:55.197Z|07361|stream_ssl|WARN|SSL_connect:
>>>>>>>>>>>>>>>>>>>>>>>>
unexpected SSL connection close
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>
Can you please compare the output of
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>
ovs-vsctl --no-wait get open .
>>>>>>>>>>>>>>>>>>>>>>>
external-ids:ovn-remote
>>>>>>>>>>>>>>>>>>>>>>>
ovs-vsctl --no-wait get open .
>>>>>>>>>>>>>>>>>>>>>>>
external-ids:ovn-encap-type
>>>>>>>>>>>>>>>>>>>>>>>
ovs-vsctl --no-wait get open .
>>>>>>>>>>>>>>>>>>>>>>>
external-ids:ovn-encap-ip
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>
of the working hosts, e.g. dc01-node02, and the
>>>>>>>>>>>>>>>>>>>>>>>
failing host dc01-node1?
>>>>>>>>>>>>>>>>>>>>>>>
This should point us the relevant difference in the
>>>>>>>>>>>>>>>>>>>>>>>
configuration.
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>
Please include ovirt-users list in your replay, to
>>>>>>>>>>>>>>>>>>>>>>>
share the knowledge and experience with the community.
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>
Thank you
>>>>>>>>>>>>>>>>>>>>>>>>
Best regards
>>>>>>>>>>>>>>>>>>>>>>>>
Konstantinos Betsis
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>
On Fri, Sep 11, 2020 at 11:01 AM Dominik Holler <
>>>>>>>>>>>>>>>>>>>>>>>>
dholler(a)redhat.com> wrote:
>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>
On Thu, Sep 10, 2020 at 6:26 PM Konstantinos B <
>>>>>>>>>>>>>>>>>>>>>>>>>
k.betsis(a)gmail.com> wrote:
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>
Hi all
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>
We have a small installation based on OVIRT 4.3.
>>>>>>>>>>>>>>>>>>>>>>>>>>
1 Cluster is based on Centos 7 and the other on
>>>>>>>>>>>>>>>>>>>>>>>>>>
OVIRT NG Node image.
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>
The environment was stable till an upgrade took
>>>>>>>>>>>>>>>>>>>>>>>>>>
place a couple of months ago.
>>>>>>>>>>>>>>>>>>>>>>>>>>
As such we had to re-install one of the Centos 7
>>>>>>>>>>>>>>>>>>>>>>>>>>
node and start from scratch.
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>
To trigger the automatic configuration of the
>>>>>>>>>>>>>>>>>>>>>>>>>
host, it is required to configure ovirt-provider-ovn as the default network
>>>>>>>>>>>>>>>>>>>>>>>>>
provider for the cluster before adding the host to oVirt.
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>
Even though the installation completed
>>>>>>>>>>>>>>>>>>>>>>>>>>
successfully and VMs are created, the following are not working as expected:
>>>>>>>>>>>>>>>>>>>>>>>>>>
1. ovn geneve tunnels are not established with
>>>>>>>>>>>>>>>>>>>>>>>>>>
the other Centos 7 node in the cluster.
>>>>>>>>>>>>>>>>>>>>>>>>>>
2. Centos 7 node is configured by ovirt engine
>>>>>>>>>>>>>>>>>>>>>>>>>>
however no geneve tunnel is established when "ovn-sbctl show" is issued on
>>>>>>>>>>>>>>>>>>>>>>>>>>
the engine.
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>
Does "ovn-sbctl show" list the hosts?
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>
3. no flows are shown on the engine on port 6642
>>>>>>>>>>>>>>>>>>>>>>>>>>
for the ovs db.
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>
Does anyone have any experience on how to
>>>>>>>>>>>>>>>>>>>>>>>>>>
troubleshoot OVN on ovirt?
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>
/var/log/openvswitch/ovncontroller.log on the
>>>>>>>>>>>>>>>>>>>>>>>>>
host should contain a helpful hint.
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>>
Thank you
>>>>>>>>>>>>>>>>>>>>>>>>>>
_______________________________________________
>>>>>>>>>>>>>>>>>>>>>>>>>>
Users mailing list -- users(a)ovirt.org
>>>>>>>>>>>>>>>>>>>>>>>>>>
To unsubscribe send an email to
>>>>>>>>>>>>>>>>>>>>>>>>>>
users-leave(a)ovirt.org
>>>>>>>>>>>>>>>>>>>>>>>>>>
Privacy Statement:
>>>>>>>>>>>>>>>>>>>>>>>>>>
https://www.ovirt.org/privacy-policy.html
>>>>>>>>>>>>>>>>>>>>>>>>>>
oVirt Code of Conduct:
>>>>>>>>>>>>>>>>>>>>>>>>>>
https://www.ovirt.org/community/about/community-guidelines/
>>>>>>>>>>>>>>>>>>>>>>>>>>
List Archives:
>>>>>>>>>>>>>>>>>>>>>>>>>>
https://lists.ovirt.org/archives/list/users@ovirt.org/message/LBVGLQJBWJF...
>>>>>>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>>>>>