Hi Dominik
Below is the output of the ovs-vsctl list interface
_uuid : bdaf92c1-4389-4ddf-aab0-93975076ebb2
admin_state : up
bfd : {}
bfd_status : {}
cfm_fault : []
cfm_fault_status : []
cfm_flap_count : []
cfm_health : []
cfm_mpid : []
cfm_remote_mpids : []
cfm_remote_opstate : []
duplex : full
error : []
external_ids : {attached-mac="56:6f:77:61:00:02",
iface-id="5d03a7a5-82a1-40f9-b50c-353a26167fa3", iface-status=active,
vm-id="e45b7b34-24f2-41ed-b95e-cd1ad532e8d3"}
ifindex : 34
ingress_policing_burst: 0
ingress_policing_rate: 0
lacp_current : []
link_resets : 1
link_speed : 10000000
link_state : up
lldp : {}
mac : []
mac_in_use : "fe:6f:77:61:00:02"
mtu : 1442
mtu_request : []
name : "vnet6"
ofport : 2
ofport_request : []
options : {}
other_config : {}
statistics : {collisions=0, rx_bytes=10828495, rx_crc_err=0,
rx_dropped=0, rx_errors=0, rx_frame_err=0, rx_over_err=0,
rx_packets=117713, tx_bytes=20771797, tx_dropped=0, tx_errors=0,
tx_packets=106954}
status : {driver_name=tun, driver_version="1.6",
firmware_version=""}
type : ""
_uuid : bad80911-3993-4085-a0b0-962b6c9156cd
admin_state : up
bfd : {}
bfd_status : {}
cfm_fault : []
cfm_fault_status : []
cfm_flap_count : []
cfm_health : []
cfm_mpid : []
cfm_remote_mpids : []
cfm_remote_opstate : []
duplex : []
error : []
external_ids : {}
ifindex : 39
ingress_policing_burst: 0
ingress_policing_rate: 0
lacp_current : []
link_resets : 0
link_speed : []
link_state : up
lldp : {}
mac : []
mac_in_use : "fe:37:52:c4:cb:03"
mtu : []
mtu_request : []
name : "ovn-c4b238-0"
ofport : 7
ofport_request : []
options : {csum="true", key=flow,
remote_ip="192.168.121.164"}
other_config : {}
statistics : {rx_bytes=0, rx_packets=0, tx_bytes=0, tx_packets=0}
status : {tunnel_egress_iface="ovirtmgmt-ams03",
tunnel_egress_iface_carrier=up}
type : geneve
_uuid : 8e7705d1-0b9d-4e30-8277-c339e7e1c27a
admin_state : up
bfd : {}
bfd_status : {}
cfm_fault : []
cfm_fault_status : []
cfm_flap_count : []
cfm_health : []
cfm_mpid : []
cfm_remote_mpids : []
cfm_remote_opstate : []
duplex : full
error : []
external_ids : {attached-mac="56:6f:77:61:00:0d",
iface-id="b7ff5f2b-4bb4-4250-8ad8-8a7e19d2b4c7", iface-status=active,
vm-id="8d73f333-bca4-4b32-9b87-2e7ee07eda84"}
ifindex : 28
ingress_policing_burst: 0
ingress_policing_rate: 0
lacp_current : []
link_resets : 1
link_speed : 10000000
link_state : up
lldp : {}
mac : []
mac_in_use : "fe:6f:77:61:00:0d"
mtu : 1442
mtu_request : []
name : "vnet0"
ofport : 1
ofport_request : []
options : {}
other_config : {}
statistics : {collisions=0, rx_bytes=20609787, rx_crc_err=0,
rx_dropped=0, rx_errors=0, rx_frame_err=0, rx_over_err=0,
rx_packets=104535, tx_bytes=10830007, tx_dropped=0, tx_errors=0,
tx_packets=117735}
status : {driver_name=tun, driver_version="1.6",
firmware_version=""}
type : ""
_uuid : 86dcc68a-63e4-4445-9373-81c1f4502c17
admin_state : up
bfd : {}
bfd_status : {}
cfm_fault : []
cfm_fault_status : []
cfm_flap_count : []
cfm_health : []
cfm_mpid : []
cfm_remote_mpids : []
cfm_remote_opstate : []
duplex : full
error : []
external_ids : {attached-mac="56:6f:77:61:00:10",
iface-id="4e8d5636-4110-41b2-906d-f9b04c2e62cd", iface-status=active,
vm-id="9a002a9b-5f09-4def-a531-d50ff683470b"}
ifindex : 40
ingress_policing_burst: 0
ingress_policing_rate: 0
lacp_current : []
link_resets : 1
link_speed : 10000000
link_state : up
lldp : {}
mac : []
mac_in_use : "fe:6f:77:61:00:10"
mtu : 1442
mtu_request : []
name : "vnet11"
ofport : 10
ofport_request : []
options : {}
other_config : {}
statistics : {collisions=0, rx_bytes=3311352, rx_crc_err=0,
rx_dropped=0, rx_errors=0, rx_frame_err=0, rx_over_err=0, rx_packets=51012,
tx_bytes=5514116, tx_dropped=0, tx_errors=0, tx_packets=103456}
status : {driver_name=tun, driver_version="1.6",
firmware_version=""}
type : ""
_uuid : e8d5e4a2-b9a0-4146-8d98-34713cb443de
admin_state : up
bfd : {}
bfd_status : {}
cfm_fault : []
cfm_fault_status : []
cfm_flap_count : []
cfm_health : []
cfm_mpid : []
cfm_remote_mpids : []
cfm_remote_opstate : []
duplex : full
error : []
external_ids : {attached-mac="56:6f:77:61:00:15",
iface-id="b88de6e4-6d77-4e42-b734-4cc676728910", iface-status=active,
vm-id="e45b7b34-24f2-41ed-b95e-cd1ad532e8d3"}
ifindex : 37
ingress_policing_burst: 0
ingress_policing_rate: 0
lacp_current : []
link_resets : 1
link_speed : 10000000
link_state : up
lldp : {}
mac : []
mac_in_use : "fe:6f:77:61:00:15"
mtu : 1442
mtu_request : []
name : "vnet9"
ofport : 5
ofport_request : []
options : {}
other_config : {}
statistics : {collisions=0, rx_bytes=180, rx_crc_err=0,
rx_dropped=0, rx_errors=0, rx_frame_err=0, rx_over_err=0, rx_packets=2,
tx_bytes=4500, tx_dropped=0, tx_errors=0, tx_packets=74}
status : {driver_name=tun, driver_version="1.6",
firmware_version=""}
type : ""
_uuid : 6a2974b3-cd72-4688-a630-0a7e9c779b21
admin_state : up
bfd : {}
bfd_status : {}
cfm_fault : []
cfm_fault_status : []
cfm_flap_count : []
cfm_health : []
cfm_mpid : []
cfm_remote_mpids : []
cfm_remote_opstate : []
duplex : full
error : []
external_ids : {attached-mac="56:6f:77:61:00:17",
iface-id="64681036-26e2-41d7-b73f-ab5302610145", iface-status=active,
vm-id="bf0dc78c-dad5-41a0-914c-ae0da0f9a388"}
ifindex : 41
ingress_policing_burst: 0
ingress_policing_rate: 0
lacp_current : []
link_resets : 1
link_speed : 10000000
link_state : up
lldp : {}
mac : []
mac_in_use : "fe:6f:77:61:00:17"
mtu : 1442
mtu_request : []
name : "vnet12"
ofport : 11
ofport_request : []
options : {}
other_config : {}
statistics : {collisions=0, rx_bytes=5513640, rx_crc_err=0,
rx_dropped=0, rx_errors=0, rx_frame_err=0, rx_over_err=0,
rx_packets=103450, tx_bytes=3311868, tx_dropped=0, tx_errors=0,
tx_packets=51018}
status : {driver_name=tun, driver_version="1.6",
firmware_version=""}
type : ""
_uuid : 44498e54-f122-41a0-a41a-7a88ba2dba9b
admin_state : down
bfd : {}
bfd_status : {}
cfm_fault : []
cfm_fault_status : []
cfm_flap_count : []
cfm_health : []
cfm_mpid : []
cfm_remote_mpids : []
cfm_remote_opstate : []
duplex : []
error : []
external_ids : {}
ifindex : 7
ingress_policing_burst: 0
ingress_policing_rate: 0
lacp_current : []
link_resets : 0
link_speed : []
link_state : down
lldp : {}
mac : []
mac_in_use : "32:0a:69:67:07:4f"
mtu : 1442
mtu_request : []
name : br-int
ofport : 65534
ofport_request : []
options : {}
other_config : {}
statistics : {collisions=0, rx_bytes=0, rx_crc_err=0,
rx_dropped=326, rx_errors=0, rx_frame_err=0, rx_over_err=0, rx_packets=0,
tx_bytes=0, tx_dropped=0, tx_errors=0, tx_packets=0}
status : {driver_name=openvswitch}
type : internal
_uuid : e2114584-8ceb-43d6-817b-e457738ead8a
admin_state : up
bfd : {}
bfd_status : {}
cfm_fault : []
cfm_fault_status : []
cfm_flap_count : []
cfm_health : []
cfm_mpid : []
cfm_remote_mpids : []
cfm_remote_opstate : []
duplex : full
error : []
external_ids : {attached-mac="56:6f:77:61:00:03",
iface-id="16162721-c815-4cd8-ab57-f22e6e482c7f", iface-status=active,
vm-id="e45b7b34-24f2-41ed-b95e-cd1ad532e8d3"}
ifindex : 35
ingress_policing_burst: 0
ingress_policing_rate: 0
lacp_current : []
link_resets : 1
link_speed : 10000000
link_state : up
lldp : {}
mac : []
mac_in_use : "fe:6f:77:61:00:03"
mtu : 1442
mtu_request : []
name : "vnet7"
ofport : 3
ofport_request : []
options : {}
other_config : {}
statistics : {collisions=0, rx_bytes=180, rx_crc_err=0,
rx_dropped=0, rx_errors=0, rx_frame_err=0, rx_over_err=0, rx_packets=2,
tx_bytes=4730, tx_dropped=0, tx_errors=0, tx_packets=77}
status : {driver_name=tun, driver_version="1.6",
firmware_version=""}
type : ""
_uuid : ee16943e-d145-4080-893f-464098a6388f
admin_state : up
bfd : {}
bfd_status : {}
cfm_fault : []
cfm_fault_status : []
cfm_flap_count : []
cfm_health : []
cfm_mpid : []
cfm_remote_mpids : []
cfm_remote_opstate : []
duplex : []
error : []
external_ids : {}
ifindex : 39
ingress_policing_burst: 0
ingress_policing_rate: 0
lacp_current : []
link_resets : 0
link_speed : []
link_state : up
lldp : {}
mac : []
mac_in_use : "1e:50:3f:a8:42:d1"
mtu : []
mtu_request : []
name : "ovn-be3abc-0"
ofport : 8
ofport_request : []
options : {csum="true", key=flow,
remote_ip="DC01-host02"}
other_config : {}
statistics : {rx_bytes=0, rx_packets=0, tx_bytes=0, tx_packets=0}
status : {tunnel_egress_iface="ovirtmgmt-ams03",
tunnel_egress_iface_carrier=up}
type : geneve
_uuid : 86a229be-373e-4c43-b2f1-6190523ed73a
admin_state : up
bfd : {}
bfd_status : {}
cfm_fault : []
cfm_fault_status : []
cfm_flap_count : []
cfm_health : []
cfm_mpid : []
cfm_remote_mpids : []
cfm_remote_opstate : []
duplex : full
error : []
external_ids : {attached-mac="56:6f:77:61:00:1c",
iface-id="12d829c3-64eb-44bc-a0bd-d7219991f35f", iface-status=active,
vm-id="e45b7b34-24f2-41ed-b95e-cd1ad532e8d3"}
ifindex : 38
ingress_policing_burst: 0
ingress_policing_rate: 0
lacp_current : []
link_resets : 1
link_speed : 10000000
link_state : up
lldp : {}
mac : []
mac_in_use : "fe:6f:77:61:00:1c"
mtu : 1442
mtu_request : []
name : "vnet10"
ofport : 6
ofport_request : []
options : {}
other_config : {}
statistics : {collisions=0, rx_bytes=117912, rx_crc_err=0,
rx_dropped=0, rx_errors=0, rx_frame_err=0, rx_over_err=0, rx_packets=2195,
tx_bytes=4204, tx_dropped=0, tx_errors=0, tx_packets=66}
status : {driver_name=tun, driver_version="1.6",
firmware_version=""}
type : ""
_uuid : fa4b8d96-bffe-4b56-930e-0e7fcc5f68ac
admin_state : up
bfd : {}
bfd_status : {}
cfm_fault : []
cfm_fault_status : []
cfm_flap_count : []
cfm_health : []
cfm_mpid : []
cfm_remote_mpids : []
cfm_remote_opstate : []
duplex : []
error : []
external_ids : {}
ifindex : 39
ingress_policing_burst: 0
ingress_policing_rate: 0
lacp_current : []
link_resets : 0
link_speed : []
link_state : up
lldp : {}
mac : []
mac_in_use : "7a:28:24:eb:ec:d2"
mtu : []
mtu_request : []
name : "ovn-95ccb0-0"
ofport : 9
ofport_request : []
options : {csum="true", key=flow,
remote_ip="DC01-host01"}
other_config : {}
statistics : {rx_bytes=0, rx_packets=0, tx_bytes=12840478,
tx_packets=224029}
status : {tunnel_egress_iface="ovirtmgmt-ams03",
tunnel_egress_iface_carrier=up}
type : geneve
_uuid : 5e3df5c7-958c-491d-8d41-0ae83c613f1d
admin_state : up
bfd : {}
bfd_status : {}
cfm_fault : []
cfm_fault_status : []
cfm_flap_count : []
cfm_health : []
cfm_mpid : []
cfm_remote_mpids : []
cfm_remote_opstate : []
duplex : full
error : []
external_ids : {attached-mac="56:6f:77:61:00:06",
iface-id="9a6cc189-0934-4468-97ae-09f90fa4598d", iface-status=active,
vm-id="e45b7b34-24f2-41ed-b95e-cd1ad532e8d3"}
ifindex : 36
ingress_policing_burst: 0
ingress_policing_rate: 0
lacp_current : []
link_resets : 1
link_speed : 10000000
link_state : up
lldp : {}
mac : []
mac_in_use : "fe:6f:77:61:00:06"
mtu : 1442
mtu_request : []
name : "vnet8"
ofport : 4
ofport_request : []
options : {}
other_config : {}
statistics : {collisions=0, rx_bytes=180, rx_crc_err=0,
rx_dropped=0, rx_errors=0, rx_frame_err=0, rx_over_err=0, rx_packets=2,
tx_bytes=8829812, tx_dropped=0, tx_errors=0, tx_packets=154540}
status : {driver_name=tun, driver_version="1.6",
firmware_version=""}
type : ""
I've identified which VMs have these MAC addresses but i do not see any
"conflict" with any other VM's MAC address.
I really do not understand why these will create a conflict.
On Wed, Sep 16, 2020 at 12:06 PM Dominik Holler <dholler(a)redhat.com> wrote:
On Tue, Sep 15, 2020 at 6:53 PM Konstantinos Betsis <k.betsis(a)gmail.com>
wrote:
> So a new test-net was created under DC01 and was depicted in the networks
> tab under both DC01 and DC02.
> I believe for some reason networks are duplicated in DCs, maybe for
> future use??? Don't know.
> If one tries to delete the network from the other DC it gets an error,
> while if deleted from the once initially created it gets deleted from both.
>
>
In oVirt a logical network is an entity in a data center. If the automatic
synchronization is enabled on the ovirt-provider-ovn entity in oVirt
Engine, the OVN networks are reflected to all data centers. If you do not
like this, you can disable the automatic synchronization of the
ovirt-provider-ovn in Admin Portal.
> From the DC01-node02 i get the following errors:
>
> 2020-09-15T16:48:49.904Z|22748|main|INFO|OVNSB commit failed, force
> recompute next time.
> 2020-09-15T16:48:49.905Z|22749|binding|INFO|Claiming lport
> 9a6cc189-0934-4468-97ae-09f90fa4598d for this chassis.
> 2020-09-15T16:48:49.905Z|22750|binding|INFO|9a6cc189-0934-4468-97ae-09f90fa4598d:
> Claiming 56:6f:77:61:00:06
> 2020-09-15T16:48:49.905Z|22751|binding|INFO|Claiming lport
> 16162721-c815-4cd8-ab57-f22e6e482c7f for this chassis.
> 2020-09-15T16:48:49.905Z|22752|binding|INFO|16162721-c815-4cd8-ab57-f22e6e482c7f:
> Claiming 56:6f:77:61:00:03
> 2020-09-15T16:48:49.905Z|22753|binding|INFO|Claiming lport
> b88de6e4-6d77-4e42-b734-4cc676728910 for this chassis.
> 2020-09-15T16:48:49.905Z|22754|binding|INFO|b88de6e4-6d77-4e42-b734-4cc676728910:
> Claiming 56:6f:77:61:00:15
> 2020-09-15T16:48:49.905Z|22755|binding|INFO|Claiming lport
> b7ff5f2b-4bb4-4250-8ad8-8a7e19d2b4c7 for this chassis.
> 2020-09-15T16:48:49.905Z|22756|binding|INFO|b7ff5f2b-4bb4-4250-8ad8-8a7e19d2b4c7:
> Claiming 56:6f:77:61:00:0d
> 2020-09-15T16:48:49.905Z|22757|binding|INFO|Claiming lport
> 5d03a7a5-82a1-40f9-b50c-353a26167fa3 for this chassis.
> 2020-09-15T16:48:49.905Z|22758|binding|INFO|5d03a7a5-82a1-40f9-b50c-353a26167fa3:
> Claiming 56:6f:77:61:00:02
> 2020-09-15T16:48:49.905Z|22759|binding|INFO|Claiming lport
> 12d829c3-64eb-44bc-a0bd-d7219991f35f for this chassis.
> 2020-09-15T16:48:49.905Z|22760|binding|INFO|12d829c3-64eb-44bc-a0bd-d7219991f35f:
> Claiming 56:6f:77:61:00:1c
> 2020-09-15T16:48:49.959Z|22761|main|INFO|OVNSB commit failed, force
> recompute next time.
> 2020-09-15T16:48:49.960Z|22762|binding|INFO|Claiming lport
> 9a6cc189-0934-4468-97ae-09f90fa4598d for this chassis.
> 2020-09-15T16:48:49.960Z|22763|binding|INFO|9a6cc189-0934-4468-97ae-09f90fa4598d:
> Claiming 56:6f:77:61:00:06
> 2020-09-15T16:48:49.960Z|22764|binding|INFO|Claiming lport
> 16162721-c815-4cd8-ab57-f22e6e482c7f for this chassis.
> 2020-09-15T16:48:49.960Z|22765|binding|INFO|16162721-c815-4cd8-ab57-f22e6e482c7f:
> Claiming 56:6f:77:61:00:03
> 2020-09-15T16:48:49.960Z|22766|binding|INFO|Claiming lport
> b88de6e4-6d77-4e42-b734-4cc676728910 for this chassis.
> 2020-09-15T16:48:49.960Z|22767|binding|INFO|b88de6e4-6d77-4e42-b734-4cc676728910:
> Claiming 56:6f:77:61:00:15
> 2020-09-15T16:48:49.960Z|22768|binding|INFO|Claiming lport
> b7ff5f2b-4bb4-4250-8ad8-8a7e19d2b4c7 for this chassis.
> 2020-09-15T16:48:49.960Z|22769|binding|INFO|b7ff5f2b-4bb4-4250-8ad8-8a7e19d2b4c7:
> Claiming 56:6f:77:61:00:0d
> 2020-09-15T16:48:49.960Z|22770|binding|INFO|Claiming lport
> 5d03a7a5-82a1-40f9-b50c-353a26167fa3 for this chassis.
> 2020-09-15T16:48:49.960Z|22771|binding|INFO|5d03a7a5-82a1-40f9-b50c-353a26167fa3:
> Claiming 56:6f:77:61:00:02
> 2020-09-15T16:48:49.960Z|22772|binding|INFO|Claiming lport
> 12d829c3-64eb-44bc-a0bd-d7219991f35f for this chassis.
> 2020-09-15T16:48:49.960Z|22773|binding|INFO|12d829c3-64eb-44bc-a0bd-d7219991f35f:
> Claiming 56:6f:77:61:00:1c
>
>
> And this repeats forever.
>
>
Looks like the southbound db is confused.
Can you try to delete all chassis listed by
sudo ovn-sbctl show
via
sudo /usr/share/ovirt-provider-ovn/scripts/remove_chassis.sh dev-host0
?
if the script remove_chassis.sh is not installed, you can use
https://github.com/oVirt/ovirt-provider-ovn/blob/master/provider/scripts/...
instead.
Can you please also share the output of
ovs-vsctl list Interface
on the host which produced the logfile above?
> The connections to ovn-sbctl is ok and the geneve tunnels are depicted
> under ovs-vsctl ok.
> VMs still not able to ping each other.
>
> On Tue, Sep 15, 2020 at 7:22 PM Dominik Holler <dholler(a)redhat.com>
> wrote:
>
>>
>>
>> On Tue, Sep 15, 2020 at 6:18 PM Konstantinos Betsis <k.betsis(a)gmail.com>
>> wrote:
>>
>>> Hi Dominik
>>>
>>> Fixed the issue.
>>>
>>
>> Thanks.
>>
>>
>>> I believe the
/etc/ovirt-provider-ovn/conf.d/10-setup-ovirt-provider-ovn.conf
>>> needed update also.
>>> The package is upgraded to the latest version.
>>>
>>> Once the provider was updated with the following it functioned
>>> perfectly:
>>>
>>> Name: ovirt-provider-ovn
>>> Description: oVirt network provider for OVN
>>> Type: External Network Provider
>>> Network Plugin: oVirt Network Provider for OVN
>>> Automatic Synchronization: Checked
>>> Unmanaged: Unchecked
>>> Provider URL: https:dc02-ovirt01.testdomain.com:9696
>>> Requires Authentication: Checked
>>> Username: admin@internal
>>> Password: "The admin password"
>>> Protocol: HTTPS
>>> Host Name:
dc02-ovirt01.testdomain.com
>>> API Port: 35357
>>> API Version: v2.0
>>> Tenant Name: "Empty"
>>>
>>> For some reason the TLS certificate was in conflict with the ovn
>>> provider details, i would bet the "host" entry.
>>>
>>> So now geneve tunnels are established.
>>> OVN provider is working.
>>>
>>> But VMs still do not communicated on the same VM network spanning
>>> different hosts.
>>>
>>> So if we have a VM network test-net on both dc01-host01 and dc01-host02
>>> and each host has a VM with IP addresses on the same network, VMs on the
>>> same VM network should communicate directly.
>>> But traffic does not reach each other.
>>>
>>>
>> Can you create a new external network, with port security disabled, and
>> an IPv4 subnet?
>> If the VMs get an IP address via DHCP, ovn is working, and should be
>> able to ping each other, too.
>> If not, there should be a helpful entry in the ovn-controller.log of the
>> host the VM is running.
>>
>>
>>> On Tue, Sep 15, 2020 at 7:07 PM Dominik Holler <dholler(a)redhat.com>
>>> wrote:
>>>
>>>> Can you try again with:
>>>>
>>>> [OVN REMOTE]
>>>> ovn-remote=ssl:127.0.0.1:6641
>>>> [SSL]
>>>> https-enabled=false
>>>> ssl-cacert-file=/etc/pki/ovirt-engine/ca.pem
>>>> ssl-cert-file=/etc/pki/ovirt-engine/certs/ovirt-provider-ovn.cer
>>>> ssl-key-file=/etc/pki/ovirt-engine/keys/ovirt-provider-ovn.key.nopass
>>>> [OVIRT]
>>>> ovirt-sso-client-secret=*random_test*
>>>> ovirt-host=https://dc02-ovirt01.testdomain.com:443
>>>> <
https://dc02-ovirt01.testdomain.com/>
>>>> ovirt-sso-client-id=ovirt-provider-ovn
>>>> ovirt-ca-file=/etc/pki/ovirt-engine/apache-ca.pem
>>>> [NETWORK]
>>>> port-security-enabled-default=True
>>>> [PROVIDER]
>>>>
>>>>
provider-host=dc02-ovirt01.testdomain.com
>>>>
>>>>
>>>>
>>>> Please note that the should match the HTTP or HTTPS in the of the
>>>> ovirt-prover-ovn configuration in oVirt Engine.
>>>> So if the ovirt-provider-ovn entity in Engine is on HTTP, the config
>>>> file should use
>>>> https-enabled=false
>>>>
>>>>
>>>> On Tue, Sep 15, 2020 at 5:56 PM Konstantinos Betsis <
>>>> k.betsis(a)gmail.com> wrote:
>>>>
>>>>> This is the updated one:
>>>>>
>>>>> # This file is automatically generated by engine-setup. Please do
not
>>>>> edit manually
>>>>> [OVN REMOTE]
>>>>> ovn-remote=ssl:127.0.0.1:6641
>>>>> [SSL]
>>>>> https-enabled=true
>>>>> ssl-cacert-file=/etc/pki/ovirt-engine/ca.pem
>>>>> ssl-cert-file=/etc/pki/ovirt-engine/certs/ovirt-provider-ovn.cer
>>>>>
ssl-key-file=/etc/pki/ovirt-engine/keys/ovirt-provider-ovn.key.nopass
>>>>> [OVIRT]
>>>>> ovirt-sso-client-secret=*random_text*
>>>>> ovirt-host=https://dc02-ovirt01.testdomain.com:443
>>>>> ovirt-sso-client-id=ovirt-provider-ovn
>>>>> ovirt-ca-file=/etc/pki/ovirt-engine/apache-ca.pem
>>>>> [NETWORK]
>>>>> port-security-enabled-default=True
>>>>> [PROVIDER]
>>>>>
provider-host=dc02-ovirt01.testdomain.com
>>>>> [AUTH]
>>>>> auth-plugin=auth.plugins.static_token:NoAuthPlugin
>>>>>
>>>>>
>>>>> However, it still does not connect.
>>>>> It prompts for the certificate but then fails and prompts to see the
>>>>> log but the ovirt-provider-ovn.log does not list anything.
>>>>>
>>>>> Yes we've got ovirt for about a year now from about version 4.1
>>>>>
>>>>>
>>>> This might explain the trouble. Upgrade of ovirt-provider-ovn should
>>>> work flawlessly starting from oVirt 4.2.
>>>>
>>>>
>>>>> On Tue, Sep 15, 2020 at 6:44 PM Dominik Holler
<dholler(a)redhat.com>
>>>>> wrote:
>>>>>
>>>>>>
>>>>>>
>>>>>> On Tue, Sep 15, 2020 at 5:34 PM Konstantinos Betsis <
>>>>>> k.betsis(a)gmail.com> wrote:
>>>>>>
>>>>>>> There is a file with the below entries
>>>>>>>
>>>>>>
>>>>>> Impressive, do you know when this config file was created and if
it
>>>>>> was manually modified?
>>>>>> Is this an upgrade from oVirt 4.1?
>>>>>>
>>>>>>
>>>>>>> [root@dc02-ovirt01 log]# cat
>>>>>>>
/etc/ovirt-provider-ovn/conf.d/10-setup-ovirt-provider-ovn.conf
>>>>>>> # This file is automatically generated by engine-setup.
Please do
>>>>>>> not edit manually
>>>>>>> [OVN REMOTE]
>>>>>>> ovn-remote=tcp:127.0.0.1:6641
>>>>>>> [SSL]
>>>>>>> https-enabled=false
>>>>>>> ssl-cacert-file=/etc/pki/ovirt-engine/ca.pem
>>>>>>>
ssl-cert-file=/etc/pki/ovirt-engine/certs/ovirt-provider-ovn.cer
>>>>>>>
>>>>>>>
ssl-key-file=/etc/pki/ovirt-engine/keys/ovirt-provider-ovn.key.nopass
>>>>>>> [OVIRT]
>>>>>>> ovirt-sso-client-secret=*random_test*
>>>>>>> ovirt-host=https://dc02-ovirt01.testdomain.com:443
>>>>>>> ovirt-sso-client-id=ovirt-provider-ovn
>>>>>>> ovirt-ca-file=/etc/pki/ovirt-engine/apache-ca.pem
>>>>>>> [NETWORK]
>>>>>>> port-security-enabled-default=True
>>>>>>> [PROVIDER]
>>>>>>>
>>>>>>>
provider-host=dc02-ovirt01.testdomain.com
>>>>>>>
>>>>>>> The only entry missing is the [AUTH] and under [SSL] the
>>>>>>> https-enabled is false. Should I edit this in this file or is
this going to
>>>>>>> break everything?
>>>>>>>
>>>>>>>
>>>>>> Changing the file should improve, but better create a backup
into
>>>>>> another diretory before modification.
>>>>>> The only required change is
>>>>>> from
>>>>>> ovn-remote=tcp:127.0.0.1:6641
>>>>>> to
>>>>>> ovn-remote=ssl:127.0.0.1:6641
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>> On Tue, Sep 15, 2020 at 6:27 PM Dominik Holler
<dholler(a)redhat.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> On Tue, Sep 15, 2020 at 5:11 PM Konstantinos Betsis <
>>>>>>>> k.betsis(a)gmail.com> wrote:
>>>>>>>>
>>>>>>>>> Hi Dominik
>>>>>>>>>
>>>>>>>>> That immediately fixed the geneve tunnels between all
hosts.
>>>>>>>>>
>>>>>>>>>
>>>>>>>> thanks for the feedback.
>>>>>>>>
>>>>>>>>
>>>>>>>>> However, the ovn provider is not broken.
>>>>>>>>> After fixing the networks we tried to move a VM to
the
>>>>>>>>> DC01-host01 so we powered it down and simply
configured it to run on
>>>>>>>>> dc01-node01.
>>>>>>>>>
>>>>>>>>> While checking the logs on the ovirt engine i noticed
the below:
>>>>>>>>> Failed to synchronize networks of Provider
ovirt-provider-ovn.
>>>>>>>>>
>>>>>>>>> The ovn-provider configure on the engine is the
below:
>>>>>>>>> Name: ovirt-provider-ovn
>>>>>>>>> Description: oVirt network provider for OVN
>>>>>>>>> Type: External Network Provider
>>>>>>>>> Network Plugin: oVirt Network Provider for OVN
>>>>>>>>> Automatic Synchronization: Checked
>>>>>>>>> Unmanaged: Unchecked
>>>>>>>>> Provider URL: http:localhost:9696
>>>>>>>>> Requires Authentication: Checked
>>>>>>>>> Username: admin@internal
>>>>>>>>> Password: "The admin password"
>>>>>>>>> Protocol: hTTP
>>>>>>>>> Host Name: dc02-ovirt01
>>>>>>>>> API Port: 35357
>>>>>>>>> API Version: v2.0
>>>>>>>>> Tenant Name: "Empty"
>>>>>>>>>
>>>>>>>>> In the past this was deleted by an engineer and
recreated as per
>>>>>>>>> the documentation, and it worked. Do we need to
update something due to the
>>>>>>>>> SSL on the ovn?
>>>>>>>>>
>>>>>>>>>
>>>>>>>> Is there a file in /etc/ovirt-provider-ovn/conf.d/ ?
>>>>>>>> engine-setup should have created one.
>>>>>>>> If the file is missing, for testing purposes, you can
create a
>>>>>>>> file
/etc/ovirt-provider-ovn/conf.d/00-setup-ovirt-provider-ovn-test.conf :
>>>>>>>> [PROVIDER]
>>>>>>>> provider-host=REPLACE_WITH_FQDN
>>>>>>>> [SSL]
>>>>>>>>
ssl-cert-file=/etc/pki/ovirt-engine/certs/ovirt-provider-ovn.cer
>>>>>>>>
>>>>>>>>
ssl-key-file=/etc/pki/ovirt-engine/keys/ovirt-provider-ovn.key.nopass
>>>>>>>> ssl-cacert-file=/etc/pki/ovirt-engine/ca.pem
>>>>>>>> https-enabled=true
>>>>>>>> [OVN REMOTE]
>>>>>>>> ovn-remote=ssl:127.0.0.1:6641
>>>>>>>> [AUTH]
>>>>>>>> auth-plugin=auth.plugins.static_token:NoAuthPlugin
>>>>>>>> [NETWORK]
>>>>>>>> port-security-enabled-default=True
>>>>>>>>
>>>>>>>> and restart the ovirt-provider-ovn service.
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>> From the ovn-provider logs the below is generated
after a service
>>>>>>>>> restart and when the start VM is triggered
>>>>>>>>>
>>>>>>>>> 2020-09-15 15:07:33,579 root Starting server
>>>>>>>>> 2020-09-15 15:07:33,579 root Version: 1.2.29-1
>>>>>>>>> 2020-09-15 15:07:33,579 root Build date:
20191217125241
>>>>>>>>> 2020-09-15 15:07:33,579 root Githash: cb5a80d
>>>>>>>>> 2020-09-15 15:08:26,582 root From:
::ffff:127.0.0.1:59980
>>>>>>>>> Request: GET /v2.0/ports
>>>>>>>>> 2020-09-15 15:08:26,582 root Could not retrieve
schema from tcp:
>>>>>>>>> 127.0.0.1:6641: Unknown error -1
>>>>>>>>> Traceback (most recent call last):
>>>>>>>>> File
"/usr/share/ovirt-provider-ovn/handlers/base_handler.py",
>>>>>>>>> line 138, in _handle_request
>>>>>>>>> method, path_parts, content
>>>>>>>>> File
>>>>>>>>>
"/usr/share/ovirt-provider-ovn/handlers/selecting_handler.py", line 175, in
>>>>>>>>> handle_request
>>>>>>>>> return self.call_response_handler(handler,
content,
>>>>>>>>> parameters)
>>>>>>>>> File
"/usr/share/ovirt-provider-ovn/handlers/neutron.py", line
>>>>>>>>> 35, in call_response_handler
>>>>>>>>> with NeutronApi() as ovn_north:
>>>>>>>>> File
"/usr/share/ovirt-provider-ovn/neutron/neutron_api.py",
>>>>>>>>> line 95, in __init__
>>>>>>>>> self.ovsidl, self.idl = ovn_connection.connect()
>>>>>>>>> File
"/usr/share/ovirt-provider-ovn/ovn_connection.py", line
>>>>>>>>> 46, in connect
>>>>>>>>> ovnconst.OVN_NORTHBOUND
>>>>>>>>> File
>>>>>>>>>
"/usr/lib/python2.7/site-packages/ovsdbapp/backend/ovs_idl/connection.py",
>>>>>>>>> line 127, in from_server
>>>>>>>>> helper =
idlutils.get_schema_helper(connection_string,
>>>>>>>>> schema_name)
>>>>>>>>> File
>>>>>>>>>
"/usr/lib/python2.7/site-packages/ovsdbapp/backend/ovs_idl/idlutils.py",
>>>>>>>>> line 128, in get_schema_helper
>>>>>>>>> 'err': os.strerror(err)})
>>>>>>>>> Exception: Could not retrieve schema from
tcp:127.0.0.1:6641:
>>>>>>>>> Unknown error -1
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> When i update the ovn provider from the GUI to have
>>>>>>>>>
https://localhost:9696/ and HTTPS as the protocol the
test fails.
>>>>>>>>>
>>>>>>>>> On Tue, Sep 15, 2020 at 5:35 PM Dominik Holler <
>>>>>>>>> dholler(a)redhat.com> wrote:
>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> On Mon, Sep 14, 2020 at 9:25 AM Konstantinos
Betsis <
>>>>>>>>>> k.betsis(a)gmail.com> wrote:
>>>>>>>>>>
>>>>>>>>>>> Hi Dominik
>>>>>>>>>>>
>>>>>>>>>>> When these commands are used on the
ovirt-engine host the
>>>>>>>>>>> output is the one depicted in your email.
>>>>>>>>>>> For your reference see also below:
>>>>>>>>>>>
>>>>>>>>>>> [root@ath01-ovirt01 certs]# ovn-nbctl
get-ssl
>>>>>>>>>>> Private key:
/etc/pki/ovirt-engine/keys/ovn-ndb.key.nopass
>>>>>>>>>>> Certificate:
/etc/pki/ovirt-engine/certs/ovn-ndb.cer
>>>>>>>>>>> CA Certificate: /etc/pki/ovirt-engine/ca.pem
>>>>>>>>>>> Bootstrap: false
>>>>>>>>>>> [root@ath01-ovirt01 certs]# ovn-nbctl
get-connection
>>>>>>>>>>> ptcp:6641
>>>>>>>>>>>
>>>>>>>>>>> [root@ath01-ovirt01 certs]# ovn-sbctl
get-ssl
>>>>>>>>>>> Private key:
/etc/pki/ovirt-engine/keys/ovn-sdb.key.nopass
>>>>>>>>>>> Certificate:
/etc/pki/ovirt-engine/certs/ovn-sdb.cer
>>>>>>>>>>> CA Certificate: /etc/pki/ovirt-engine/ca.pem
>>>>>>>>>>> Bootstrap: false
>>>>>>>>>>> [root@ath01-ovirt01 certs]# ovn-sbctl
get-connection
>>>>>>>>>>> read-write role="" ptcp:6642
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>> ^^^ the line above points to the problem:
ovn-central is
>>>>>>>>>> configured to use plain TCP without ssl.
>>>>>>>>>> engine-setup usually configures ovn-central to
use SSL. That the
>>>>>>>>>> files /etc/pki/ovirt-engine/keys/ovn-* exist,
shows,
>>>>>>>>>> that engine-setup was triggered correctly. Looks
like the ovn db
>>>>>>>>>> was dropped somehow, this should not happen.
>>>>>>>>>> This can be fixed manually by executing the
following commands
>>>>>>>>>> on engine's machine:
>>>>>>>>>> ovn-nbctl set-ssl
/etc/pki/ovirt-engine/keys/ovn-ndb.key.nopass
>>>>>>>>>> /etc/pki/ovirt-engine/certs/ovn-ndb.cer
/etc/pki/ovirt-engine/ca.pem
>>>>>>>>>> ovn-nbctl set-connection pssl:6641
>>>>>>>>>> ovn-sbctl set-ssl
/etc/pki/ovirt-engine/keys/ovn-sdb.key.nopass
>>>>>>>>>> /etc/pki/ovirt-engine/certs/ovn-sdb.cer
/etc/pki/ovirt-engine/ca.pem
>>>>>>>>>> ovn-sbctl set-connection pssl:6642
>>>>>>>>>>
>>>>>>>>>> The /var/log/openvswitch/ovn-controller.log on
the hosts should
>>>>>>>>>> tell that br-int.mgmt is connected now.
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>> [root@ath01-ovirt01 certs]# ls -l
>>>>>>>>>>> /etc/pki/ovirt-engine/keys/ovn-*
>>>>>>>>>>> -rw-r-----. 1 root hugetlbfs 1828 Jun 25
11:08
>>>>>>>>>>>
/etc/pki/ovirt-engine/keys/ovn-ndb.key.nopass
>>>>>>>>>>> -rw-------. 1 root root 2893 Jun 25
11:08
>>>>>>>>>>> /etc/pki/ovirt-engine/keys/ovn-ndb.p12
>>>>>>>>>>> -rw-r-----. 1 root hugetlbfs 1828 Jun 25
11:08
>>>>>>>>>>>
/etc/pki/ovirt-engine/keys/ovn-sdb.key.nopass
>>>>>>>>>>> -rw-------. 1 root root 2893 Jun 25
11:08
>>>>>>>>>>> /etc/pki/ovirt-engine/keys/ovn-sdb.p12
>>>>>>>>>>>
>>>>>>>>>>> When i try the above commands on the node
hosts the following
>>>>>>>>>>> happens:
>>>>>>>>>>> ovn-nbctl get-ssl / get-connection
>>>>>>>>>>> ovn-nbctl:
unix:/var/run/openvswitch/ovnnb_db.sock: database
>>>>>>>>>>> connection failed (No such file or
directory)
>>>>>>>>>>> The above i believe is expected since no
northbound connections
>>>>>>>>>>> should be established from the host nodes.
>>>>>>>>>>>
>>>>>>>>>>> ovn-sbctl get-ssl /get-connection
>>>>>>>>>>> The output is stuck till i terminate it.
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>> Yes, the ovn-* commands works only on
engine's machine, which
>>>>>>>>>> has the role ovn-central.
>>>>>>>>>> On the hosts, there is only the ovn-controller,
which connects
>>>>>>>>>> the ovn southbound to openvswitch on the host.
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>> For the requested logs the below are found in
the
>>>>>>>>>>> ovsdb-server-sb.log
>>>>>>>>>>>
>>>>>>>>>>>
2020-09-14T07:18:38.187Z|219636|reconnect|WARN|tcp:DC02-host01:33146:
>>>>>>>>>>> connection dropped (Protocol error)
>>>>>>>>>>>
2020-09-14T07:18:41.946Z|219637|reconnect|WARN|tcp:DC01-host01:51188:
>>>>>>>>>>> connection dropped (Protocol error)
>>>>>>>>>>>
2020-09-14T07:18:43.033Z|219638|reconnect|WARN|tcp:DC01-host02:37044:
>>>>>>>>>>> connection dropped (Protocol error)
>>>>>>>>>>>
2020-09-14T07:18:46.198Z|219639|reconnect|WARN|tcp:DC02-host01:33148:
>>>>>>>>>>> connection dropped (Protocol error)
>>>>>>>>>>>
2020-09-14T07:18:50.069Z|219640|jsonrpc|WARN|Dropped 4 log
>>>>>>>>>>> messages in last 12 seconds (most recently, 4
seconds ago) due to excessive
>>>>>>>>>>> rate
>>>>>>>>>>>
2020-09-14T07:18:50.069Z|219641|jsonrpc|WARN|tcp:DC01-host01:51190:
>>>>>>>>>>> error parsing stream: line 0, column 0, byte
0: invalid character U+0016
>>>>>>>>>>>
2020-09-14T07:18:50.069Z|219642|jsonrpc|WARN|Dropped 4 log
>>>>>>>>>>> messages in last 12 seconds (most recently, 4
seconds ago) due to excessive
>>>>>>>>>>> rate
>>>>>>>>>>>
2020-09-14T07:18:50.069Z|219643|jsonrpc|WARN|tcp:DC01-host01:51190:
>>>>>>>>>>> received SSL data on JSON-RPC channel
>>>>>>>>>>>
2020-09-14T07:18:50.070Z|219644|reconnect|WARN|tcp:DC01-host01:51190:
>>>>>>>>>>> connection dropped (Protocol error)
>>>>>>>>>>>
2020-09-14T07:18:51.147Z|219645|reconnect|WARN|tcp:DC01-host02:37046:
>>>>>>>>>>> connection dropped (Protocol error)
>>>>>>>>>>>
2020-09-14T07:18:54.209Z|219646|reconnect|WARN|tcp:DC02-host01:33150:
>>>>>>>>>>> connection dropped (Protocol error)
>>>>>>>>>>>
2020-09-14T07:18:58.192Z|219647|reconnect|WARN|tcp:DC01-host01:51192:
>>>>>>>>>>> connection dropped (Protocol error)
>>>>>>>>>>>
2020-09-14T07:18:59.262Z|219648|jsonrpc|WARN|Dropped 3 log
>>>>>>>>>>> messages in last 8 seconds (most recently, 1
seconds ago) due to excessive
>>>>>>>>>>> rate
>>>>>>>>>>>
2020-09-14T07:18:59.262Z|219649|jsonrpc|WARN|tcp:DC01-host02:37048:
>>>>>>>>>>> error parsing stream: line 0, column 0, byte
0: invalid character U+0016
>>>>>>>>>>>
2020-09-14T07:18:59.263Z|219650|jsonrpc|WARN|Dropped 3 log
>>>>>>>>>>> messages in last 8 seconds (most recently, 1
seconds ago) due to excessive
>>>>>>>>>>> rate
>>>>>>>>>>>
2020-09-14T07:18:59.263Z|219651|jsonrpc|WARN|tcp:DC01-host02:37048:
>>>>>>>>>>> received SSL data on JSON-RPC channel
>>>>>>>>>>>
2020-09-14T07:18:59.263Z|219652|reconnect|WARN|tcp:DC01-host02:37048:
>>>>>>>>>>> connection dropped (Protocol error)
>>>>>>>>>>>
2020-09-14T07:19:02.220Z|219653|reconnect|WARN|tcp:DC02-host01:33152:
>>>>>>>>>>> connection dropped (Protocol error)
>>>>>>>>>>>
2020-09-14T07:19:06.316Z|219654|reconnect|WARN|tcp:DC01-host01:51194:
>>>>>>>>>>> connection dropped (Protocol error)
>>>>>>>>>>>
2020-09-14T07:19:07.386Z|219655|reconnect|WARN|tcp:DC01-host02:37050:
>>>>>>>>>>> connection dropped (Protocol error)
>>>>>>>>>>>
2020-09-14T07:19:10.232Z|219656|reconnect|WARN|tcp:DC02-host01:33154:
>>>>>>>>>>> connection dropped (Protocol error)
>>>>>>>>>>>
2020-09-14T07:19:14.439Z|219657|jsonrpc|WARN|Dropped 4 log
>>>>>>>>>>> messages in last 12 seconds (most recently, 4
seconds ago) due to excessive
>>>>>>>>>>> rate
>>>>>>>>>>>
2020-09-14T07:19:14.439Z|219658|jsonrpc|WARN|tcp:DC01-host01:51196:
>>>>>>>>>>> error parsing stream: line 0, column 0, byte
0: invalid character U+0016
>>>>>>>>>>>
2020-09-14T07:19:14.439Z|219659|jsonrpc|WARN|Dropped 4 log
>>>>>>>>>>> messages in last 12 seconds (most recently, 4
seconds ago) due to excessive
>>>>>>>>>>> rate
>>>>>>>>>>>
2020-09-14T07:19:14.439Z|219660|jsonrpc|WARN|tcp:DC01-host01:51196:
>>>>>>>>>>> received SSL data on JSON-RPC channel
>>>>>>>>>>>
2020-09-14T07:19:14.440Z|219661|reconnect|WARN|tcp:DC01-host01:51196:
>>>>>>>>>>> connection dropped (Protocol error)
>>>>>>>>>>>
2020-09-14T07:19:15.505Z|219662|reconnect|WARN|tcp:DC01-host02:37052:
>>>>>>>>>>> connection dropped (Protocol error)
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> How can we fix these SSL errors?
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> I addressed this above.
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>> I thought vdsm did the certificate
provisioning on the host
>>>>>>>>>>> nodes as to communicate to the engine host
node.
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>> Yes, this seems to work in your scenario, just
the SSL
>>>>>>>>>> configuration on the ovn-central was lost.
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>> On Fri, Sep 11, 2020 at 6:39 PM Dominik
Holler <
>>>>>>>>>>> dholler(a)redhat.com> wrote:
>>>>>>>>>>>
>>>>>>>>>>>> Looks still like the ovn-controller on
the host has problems
>>>>>>>>>>>> communicating with ovn-southbound.
>>>>>>>>>>>>
>>>>>>>>>>>> Are there any hints in
/var/log/openvswitch/*.log,
>>>>>>>>>>>> especially in
/var/log/openvswitch/ovsdb-server-sb.log ?
>>>>>>>>>>>>
>>>>>>>>>>>> Can you please check the output of
>>>>>>>>>>>>
>>>>>>>>>>>> ovn-nbctl get-ssl
>>>>>>>>>>>> ovn-nbctl get-connection
>>>>>>>>>>>> ovn-sbctl get-ssl
>>>>>>>>>>>> ovn-sbctl get-connection
>>>>>>>>>>>> ls -l /etc/pki/ovirt-engine/keys/ovn-*
>>>>>>>>>>>>
>>>>>>>>>>>> it should be similar to
>>>>>>>>>>>>
>>>>>>>>>>>> [root@ovirt-43 ~]# ovn-nbctl get-ssl
>>>>>>>>>>>> Private key:
/etc/pki/ovirt-engine/keys/ovn-ndb.key.nopass
>>>>>>>>>>>> Certificate:
/etc/pki/ovirt-engine/certs/ovn-ndb.cer
>>>>>>>>>>>> CA Certificate:
/etc/pki/ovirt-engine/ca.pem
>>>>>>>>>>>> Bootstrap: false
>>>>>>>>>>>> [root@ovirt-43 ~]# ovn-nbctl
get-connection
>>>>>>>>>>>> pssl:6641:[::]
>>>>>>>>>>>> [root@ovirt-43 ~]# ovn-sbctl get-ssl
>>>>>>>>>>>> Private key:
/etc/pki/ovirt-engine/keys/ovn-sdb.key.nopass
>>>>>>>>>>>> Certificate:
/etc/pki/ovirt-engine/certs/ovn-sdb.cer
>>>>>>>>>>>> CA Certificate:
/etc/pki/ovirt-engine/ca.pem
>>>>>>>>>>>> Bootstrap: false
>>>>>>>>>>>> [root@ovirt-43 ~]# ovn-sbctl
get-connection
>>>>>>>>>>>> read-write role=""
pssl:6642:[::]
>>>>>>>>>>>> [root@ovirt-43 ~]# ls -l
/etc/pki/ovirt-engine/keys/ovn-*
>>>>>>>>>>>> -rw-r-----. 1 root hugetlbfs 1828 Oct 14
2019
>>>>>>>>>>>>
/etc/pki/ovirt-engine/keys/ovn-ndb.key.nopass
>>>>>>>>>>>> -rw-------. 1 root root 2709 Oct 14
2019
>>>>>>>>>>>> /etc/pki/ovirt-engine/keys/ovn-ndb.p12
>>>>>>>>>>>> -rw-r-----. 1 root hugetlbfs 1828 Oct 14
2019
>>>>>>>>>>>>
/etc/pki/ovirt-engine/keys/ovn-sdb.key.nopass
>>>>>>>>>>>> -rw-------. 1 root root 2709 Oct 14
2019
>>>>>>>>>>>> /etc/pki/ovirt-engine/keys/ovn-sdb.p12
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> On Fri, Sep 11, 2020 at 1:10 PM
Konstantinos Betsis <
>>>>>>>>>>>> k.betsis(a)gmail.com> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>>> I did a restart of the
ovn-controller, this is the output of
>>>>>>>>>>>>> the ovn-controller.log
>>>>>>>>>>>>>
>>>>>>>>>>>>>
2020-09-11T10:54:07.566Z|00001|vlog|INFO|opened log file
>>>>>>>>>>>>>
/var/log/openvswitch/ovn-controller.log
>>>>>>>>>>>>>
2020-09-11T10:54:07.568Z|00002|reconnect|INFO|unix:/var/run/openvswitch/db.sock:
>>>>>>>>>>>>> connecting...
>>>>>>>>>>>>>
2020-09-11T10:54:07.568Z|00003|reconnect|INFO|unix:/var/run/openvswitch/db.sock:
>>>>>>>>>>>>> connected
>>>>>>>>>>>>>
2020-09-11T10:54:07.570Z|00004|main|INFO|OVS IDL reconnected,
>>>>>>>>>>>>> force recompute.
>>>>>>>>>>>>>
2020-09-11T10:54:07.571Z|00005|reconnect|INFO|ssl:OVIRT_ENGINE_IP:6642:
>>>>>>>>>>>>> connecting...
>>>>>>>>>>>>>
2020-09-11T10:54:07.571Z|00006|main|INFO|OVNSB IDL
>>>>>>>>>>>>> reconnected, force recompute.
>>>>>>>>>>>>>
2020-09-11T10:54:07.685Z|00007|stream_ssl|WARN|SSL_connect:
>>>>>>>>>>>>> unexpected SSL connection close
>>>>>>>>>>>>>
2020-09-11T10:54:07.685Z|00008|reconnect|INFO|ssl:OVIRT_ENGINE_IP:6642:
>>>>>>>>>>>>> connection attempt failed (Protocol
error)
>>>>>>>>>>>>>
2020-09-11T10:54:08.685Z|00009|reconnect|INFO|ssl:OVIRT_ENGINE_IP:6642:
>>>>>>>>>>>>> connecting...
>>>>>>>>>>>>>
2020-09-11T10:54:08.800Z|00010|stream_ssl|WARN|SSL_connect:
>>>>>>>>>>>>> unexpected SSL connection close
>>>>>>>>>>>>>
2020-09-11T10:54:08.800Z|00011|reconnect|INFO|ssl:OVIRT_ENGINE_IP:6642:
>>>>>>>>>>>>> connection attempt failed (Protocol
error)
>>>>>>>>>>>>>
2020-09-11T10:54:08.800Z|00012|reconnect|INFO|ssl:OVIRT_ENGINE_IP:6642:
>>>>>>>>>>>>> waiting 2 seconds before reconnect
>>>>>>>>>>>>>
2020-09-11T10:54:10.802Z|00013|reconnect|INFO|ssl:OVIRT_ENGINE_IP:6642:
>>>>>>>>>>>>> connecting...
>>>>>>>>>>>>>
2020-09-11T10:54:10.917Z|00014|stream_ssl|WARN|SSL_connect:
>>>>>>>>>>>>> unexpected SSL connection close
>>>>>>>>>>>>>
2020-09-11T10:54:10.917Z|00015|reconnect|INFO|ssl:OVIRT_ENGINE_IP:6642:
>>>>>>>>>>>>> connection attempt failed (Protocol
error)
>>>>>>>>>>>>>
2020-09-11T10:54:10.917Z|00016|reconnect|INFO|ssl:OVIRT_ENGINE_IP:6642:
>>>>>>>>>>>>> waiting 4 seconds before reconnect
>>>>>>>>>>>>>
2020-09-11T10:54:14.921Z|00017|reconnect|INFO|ssl:OVIRT_ENGINE_IP:6642:
>>>>>>>>>>>>> connecting...
>>>>>>>>>>>>>
2020-09-11T10:54:15.036Z|00018|stream_ssl|WARN|SSL_connect:
>>>>>>>>>>>>> unexpected SSL connection close
>>>>>>>>>>>>>
2020-09-11T10:54:15.036Z|00019|reconnect|INFO|ssl:OVIRT_ENGINE_IP:6642:
>>>>>>>>>>>>> connection attempt failed (Protocol
error)
>>>>>>>>>>>>>
2020-09-11T10:54:15.036Z|00020|reconnect|INFO|ssl:OVIRT_ENGINE_IP:6642:
>>>>>>>>>>>>> continuing to reconnect in the
background but suppressing further logging
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> I have also done the vdsm-tool
ovn-config OVIRT_ENGINE_IP
>>>>>>>>>>>>> OVIRTMGMT_NETWORK_DC
>>>>>>>>>>>>> This is how the OVIRT_ENGINE_IP is
provided in the ovn
>>>>>>>>>>>>> controller, i can redo it if you
wan.
>>>>>>>>>>>>>
>>>>>>>>>>>>> After the restart of the
ovn-controller the OVIRT ENGINE
>>>>>>>>>>>>> still shows only two geneve
connections one with DC01-host02 and
>>>>>>>>>>>>> DC02-host01.
>>>>>>>>>>>>> Chassis
"c4b23834-aec7-4bf8-8be7-aa94a50a6144"
>>>>>>>>>>>>> hostname:
"dc02-host01"
>>>>>>>>>>>>> Encap geneve
>>>>>>>>>>>>> ip:
"DC02-host01_IP"
>>>>>>>>>>>>> options:
{csum="true"}
>>>>>>>>>>>>> Chassis
"be3abcc9-7358-4040-a37b-8d8a782f239c"
>>>>>>>>>>>>> hostname:
"DC01-host02"
>>>>>>>>>>>>> Encap geneve
>>>>>>>>>>>>> ip: "DC01-host02"
>>>>>>>>>>>>> options:
{csum="true"}
>>>>>>>>>>>>>
>>>>>>>>>>>>> I've re-done the vdsm-tool
command and nothing changed....
>>>>>>>>>>>>> again....with the same errors as the
systemctl restart ovn-controller
>>>>>>>>>>>>>
>>>>>>>>>>>>> On Fri, Sep 11, 2020 at 1:49 PM
Dominik Holler <
>>>>>>>>>>>>> dholler(a)redhat.com> wrote:
>>>>>>>>>>>>>
>>>>>>>>>>>>>> Please include ovirt-users list
in your reply, to share
>>>>>>>>>>>>>> the knowledge and experience with
the community!
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> On Fri, Sep 11, 2020 at 12:12 PM
Konstantinos Betsis <
>>>>>>>>>>>>>> k.betsis(a)gmail.com> wrote:
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Ok below the output per node
and DC
>>>>>>>>>>>>>>> DC01
>>>>>>>>>>>>>>> node01
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> [root@dc01-node01 ~]#
ovs-vsctl --no-wait get open .
>>>>>>>>>>>>>>> external-ids:ovn-remote
>>>>>>>>>>>>>>>
"ssl:*OVIRT_ENGINE_IP*:6642"
>>>>>>>>>>>>>>> [root@ dc01-node01 ~]#
ovs-vsctl --no-wait get open .
>>>>>>>>>>>>>>> external-ids:ovn-encap-type
>>>>>>>>>>>>>>> geneve
>>>>>>>>>>>>>>> [root@ dc01-node01 ~]#
ovs-vsctl --no-wait get open .
>>>>>>>>>>>>>>> external-ids:ovn-encap-ip
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
"*OVIRTMGMT_IP_DC01-NODE01*"
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> node02
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> [root@dc01-node02 ~]#
ovs-vsctl --no-wait get open .
>>>>>>>>>>>>>>> external-ids:ovn-remote
>>>>>>>>>>>>>>>
"ssl:*OVIRT_ENGINE_IP*:6642"
>>>>>>>>>>>>>>> [root@ dc01-node02 ~]#
ovs-vsctl --no-wait get open .
>>>>>>>>>>>>>>> external-ids:ovn-encap-type
>>>>>>>>>>>>>>> geneve
>>>>>>>>>>>>>>> [root@ dc01-node02 ~]#
ovs-vsctl --no-wait get open .
>>>>>>>>>>>>>>> external-ids:ovn-encap-ip
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
"*OVIRTMGMT_IP_DC01-NODE02*"
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> DC02
>>>>>>>>>>>>>>> node01
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> [root@dc02-node01 ~]#
ovs-vsctl --no-wait get open .
>>>>>>>>>>>>>>> external-ids:ovn-remote
>>>>>>>>>>>>>>>
"ssl:*OVIRT_ENGINE_IP*:6642"
>>>>>>>>>>>>>>> [root@ dc02-node01 ~]#
ovs-vsctl --no-wait get open .
>>>>>>>>>>>>>>> external-ids:ovn-encap-type
>>>>>>>>>>>>>>> geneve
>>>>>>>>>>>>>>> [root@ dc02-node01 ~]#
ovs-vsctl --no-wait get open .
>>>>>>>>>>>>>>> external-ids:ovn-encap-ip
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
"*OVIRTMGMT_IP_DC02-NODE01*"
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Looks good.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> DC01 node01 and node02 share
the same VM networks and VMs
>>>>>>>>>>>>>>> deployed on top of them
cannot talk to VM on the other hypervisor.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Maybe there is a hint on
ovn-controller.log on dc01-node02 ?
>>>>>>>>>>>>>> Maybe restarting ovn-controller
creates more helpful log messages?
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> You can also try restart the ovn
configuration on all hosts
>>>>>>>>>>>>>> by executing
>>>>>>>>>>>>>> vdsm-tool ovn-config
OVIRT_ENGINE_IP LOCAL_OVIRTMGMT_IP
>>>>>>>>>>>>>> on each host, this would trigger
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
https://github.com/oVirt/ovirt-provider-ovn/blob/master/driver/scripts/se...
>>>>>>>>>>>>>> internally.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> So I would expect to see the
same output for node01 to have
>>>>>>>>>>>>>>> a geneve tunnel to node02 and
vice versa.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Me too.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> On Fri, Sep 11, 2020 at 12:14
PM Dominik Holler <
>>>>>>>>>>>>>>> dholler(a)redhat.com>
wrote:
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> On Fri, Sep 11, 2020 at
10:53 AM Konstantinos Betsis <
>>>>>>>>>>>>>>>> k.betsis(a)gmail.com>
wrote:
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Hi Dominik
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> OVN is selected as
the default network provider on the
>>>>>>>>>>>>>>>>> clusters and the
hosts.
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> sounds good.
>>>>>>>>>>>>>>>> This configuration is
required already during the host is
>>>>>>>>>>>>>>>> added to oVirt Engine,
because OVN is configured during this step.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> The "ovn-sbctl
show" works on the ovirt engine and shows
>>>>>>>>>>>>>>>>> only two hosts, 1 per
DC.
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Chassis
"c4b23834-aec7-4bf8-8be7-aa94a50a6144"
>>>>>>>>>>>>>>>>> hostname:
"dc01-node02"
>>>>>>>>>>>>>>>>> Encap geneve
>>>>>>>>>>>>>>>>> ip:
"X.X.X.X"
>>>>>>>>>>>>>>>>> options:
{csum="true"}
>>>>>>>>>>>>>>>>> Chassis
"be3abcc9-7358-4040-a37b-8d8a782f239c"
>>>>>>>>>>>>>>>>> hostname:
"dc02-node1"
>>>>>>>>>>>>>>>>> Encap geneve
>>>>>>>>>>>>>>>>> ip:
"A.A.A.A"
>>>>>>>>>>>>>>>>> options:
{csum="true"}
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> The new node is not
listed (dc01-node1).
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> When executed on the
nodes the same command (ovn-sbctl
>>>>>>>>>>>>>>>>> show) times-out on
all nodes.....
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> The output of the
>>>>>>>>>>>>>>>>>
/var/log/openvswitch/ovn-conntroller.log lists on all logs
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
2020-09-11T08:46:55.197Z|07361|stream_ssl|WARN|SSL_connect:
>>>>>>>>>>>>>>>>> unexpected SSL
connection close
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Can you please compare
the output of
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> ovs-vsctl --no-wait get
open . external-ids:ovn-remote
>>>>>>>>>>>>>>>> ovs-vsctl --no-wait get
open . external-ids:ovn-encap-type
>>>>>>>>>>>>>>>> ovs-vsctl --no-wait get
open . external-ids:ovn-encap-ip
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> of the working hosts,
e.g. dc01-node02, and the failing
>>>>>>>>>>>>>>>> host dc01-node1?
>>>>>>>>>>>>>>>> This should point us the
relevant difference in the
>>>>>>>>>>>>>>>> configuration.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Please include
ovirt-users list in your replay, to share
>>>>>>>>>>>>>>>> the knowledge and
experience with the community.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Thank you
>>>>>>>>>>>>>>>>> Best regards
>>>>>>>>>>>>>>>>> Konstantinos Betsis
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> On Fri, Sep 11, 2020
at 11:01 AM Dominik Holler <
>>>>>>>>>>>>>>>>>
dholler(a)redhat.com> wrote:
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> On Thu, Sep 10,
2020 at 6:26 PM Konstantinos B <
>>>>>>>>>>>>>>>>>>
k.betsis(a)gmail.com> wrote:
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Hi all
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> We have a
small installation based on OVIRT 4.3.
>>>>>>>>>>>>>>>>>>> 1 Cluster is
based on Centos 7 and the other on OVIRT
>>>>>>>>>>>>>>>>>>> NG Node
image.
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> The
environment was stable till an upgrade took place a
>>>>>>>>>>>>>>>>>>> couple of
months ago.
>>>>>>>>>>>>>>>>>>> As such we
had to re-install one of the Centos 7 node
>>>>>>>>>>>>>>>>>>> and start
from scratch.
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> To trigger the
automatic configuration of the host, it
>>>>>>>>>>>>>>>>>> is required to
configure ovirt-provider-ovn as the default network provider
>>>>>>>>>>>>>>>>>> for the cluster
before adding the host to oVirt.
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Even though
the installation completed successfully and
>>>>>>>>>>>>>>>>>>> VMs are
created, the following are not working as expected:
>>>>>>>>>>>>>>>>>>> 1. ovn geneve
tunnels are not established with the
>>>>>>>>>>>>>>>>>>> other Centos
7 node in the cluster.
>>>>>>>>>>>>>>>>>>> 2. Centos 7
node is configured by ovirt engine however
>>>>>>>>>>>>>>>>>>> no geneve
tunnel is established when "ovn-sbctl show" is issued on the
>>>>>>>>>>>>>>>>>>> engine.
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Does
"ovn-sbctl show" list the hosts?
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> 3. no flows
are shown on the engine on port 6642 for
>>>>>>>>>>>>>>>>>>> the ovs db.
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Does anyone
have any experience on how to troubleshoot
>>>>>>>>>>>>>>>>>>> OVN on
ovirt?
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
/var/log/openvswitch/ovncontroller.log on the host
>>>>>>>>>>>>>>>>>> should contain a
helpful hint.
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Thank you
>>>>>>>>>>>>>>>>>>>
_______________________________________________
>>>>>>>>>>>>>>>>>>> Users mailing
list -- users(a)ovirt.org
>>>>>>>>>>>>>>>>>>> To
unsubscribe send an email to users-leave(a)ovirt.org
>>>>>>>>>>>>>>>>>>> Privacy
Statement:
>>>>>>>>>>>>>>>>>>>
https://www.ovirt.org/privacy-policy.html
>>>>>>>>>>>>>>>>>>> oVirt Code of
Conduct:
>>>>>>>>>>>>>>>>>>>
https://www.ovirt.org/community/about/community-guidelines/
>>>>>>>>>>>>>>>>>>> List
Archives:
>>>>>>>>>>>>>>>>>>>
https://lists.ovirt.org/archives/list/users@ovirt.org/message/LBVGLQJBWJF...
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>