Hi Miguel

if i do ovs-vsctl  list Open_vSwitch i get

uuid               : ce94c4b1-7eb2-42e3-8bfd-96e1dec40dea
bridges             : [9b0738ee-594d-4a87-8967-049a8b1a5774]
cur_cfg             : 1
datapath_types      : [netdev, system]
db_version          : "7.14.0"
external_ids        : {hostname="ovhost2", ovn-bridge-mappings="", ovn-encap-ip="", ovn-encap-type=geneve, ovn-remote="ssl:", system-id="7c39d07b-1d54-417b-bf56-7a0f1a07f832"}
iface_types         : [geneve, gre, internal, lisp, patch, stt, system, tap, vxlan]
manager_options     : []
next_cfg            : 1
other_config        : {}
ovs_version         : "2.7.3"
ssl                 : []
statistics          : {}
system_type         : centos
system_version      : "7"

I can see two addresses that are on the old network..

On Wed, Jul 17, 2019 at 8:21 AM carl langlois <crl.langlois@gmail.com> wrote:
Hi Miguel,

I will surely open a bugs, any specific ovirt componenent to select when openeing the bug?

When you say that the hosted-engine should have trigger a the update. Do you mean is was suppose to trigger the update and did not work or it is something missing?
Could i have missed a step when switching the network?

Also if i try to do ovs-vsctl list . The list command require a Table name. Not sure what table to use?


On Wed, Jul 17, 2019 at 4:21 AM Miguel Duarte de Mora Barroso <mdbarroso@redhat.com> wrote:
On Tue, Jul 16, 2019 at 8:48 PM carl langlois <crl.langlois@gmail.com> wrote:
> Hi
> We are in a process of changing our network connection. Our current network is using 10.8.256.x and we will change to 10.16.248.x. We have a HA ovirt cluster (around 10 nodes) currently configure on the 10.8.256.x. So my question is is it possible to relocate the ovirt cluster to the 10.16.248.x.  We have tried to move everything to the new network without success. All the node seem to boot up properly, our gluster storage also work properly.
> When we try to start the hosted-engine it goes up but fail the liveliness check. We have notice in the /var/log/openvswitch/ovn-controller.log that he is triying to connect to the hold ip address of the hosted-engine vm.
> 019-07-16T18:41:29.483Z|01992|reconnect|INFO|ssl: waiting 8 seconds before reconnect
> 2019-07-16T18:41:37.489Z|01993|reconnect|INFO|ssl: connecting...
> 2019-07-16T18:41:45.497Z|01994|reconnect|INFO|ssl: connection attempt timed out
> So my question is were is the come from.

Looks like the ovn controllers were not updated during the network change.

The wrong IP is configured within openvswitch, you can see it in the
(offending) nodes through "ovs-vsctl list . ". It'll be a key in the
'external_ids' column called 'ovn-remote' .

This is not the solution, but a work-around; you could try to
configure the ovn controllers via:
vdsm-tool ovn-config <engine_ip_on_net> <name of the management network>

Despite the provided work-around, I really think the hosted engine
should have triggered the ansible role that in turn triggers this

Would you open a bug with this information ?

> The routing table for one of our host look like this
> estination     Gateway         Genmask         Flags Metric Ref    Use Iface
> default         gateway         UG    0      0        0 ovirtmgmt
>   U     0      0        0 ovirtmgmt
> link-local     U     1002   0        0 eno1
> link-local     U     1003   0        0 eno2
> link-local     U     1025   0        0 ovirtmgmt
> Any help would be really appreciated.
> Regards
> Carl
> _______________________________________________
> Users mailing list -- users@ovirt.org
> To unsubscribe send an email to users-leave@ovirt.org
> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/
> List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/DBQUWEPPDK2JDFU4HOGNURK7AB3FDINC/