On Thu, Mar 17, 2022 at 11:43 AM Renaud RAKOTOMALALA <renaud.rakotomalala@alterway.fr> wrote:
Hi Ales,

Le mer. 16 mars 2022 à 07:11, Ales Musil <amusil@redhat.com> a écrit :

[../..]

I am trying to add a new ovirt-node-ng 4.4.10 node to my cluster managed by an ovirt-engine version 4.4.10.

My cluster is composed of other ovirt-node-ng which have been successively updated from version 4.4.4 to version 4.4.10 without any problem.

This new node is integrated normally in the cluster, however when I look at the status of the network part in the tab "Network interface" I see that all interfaces are "down".

Did you try to call "Refresh Capabilities"? It might be the case that the engine presents a different state that is on the host after upgrade.

I tried but and I show the pull in the vdsm.log on my faulty node. But the bond/interfaces states still "down". I tried a fresh install several time  the node with "ovirt-node-ng-installer-4.4.10-2022030308.el8.iso" but the issue still there.

 

I have a paperclip at the "bond0" interface that says: "Bond state; NONE"

I compared the content of "/etc/sysconfig/network-script" between an hypervisor which works and the one which has the problem and I notice that a whole bunch of files are missing and in particular the "ifup/ifdown...." files. The folder contains only the cluster specific files + the "ovirtmgmt" interface.

Since 4.4 in general we don't use initscripts anymore, so those files are really not a good indicator of anything. We are using nmstate + NetworkManager, if the connection are correctly presented here
everything should be fine.
 

networkManager show interfaces and bond up and running from the node perspective
nmcli con show --active
NAME       UUID                                  TYPE      DEVICE    
ovirtmgmt  6b08c819-6091-44de-9546-X  bridge    ovirtmgmt
virbr0     91cb9d5c-b64d-4655-ac2a-X  bridge    virbr0    
bond0      ad33d8b0-1f7b-cab9-9447-X  bond      bond0    
eno1       abf4c85b-57cc-4484-4fa9-X  ethernet  eno1      
eno2       b186f945-cc80-911d-668c-X  ethernet  eno2 
 
nmstatectl show return correct states

- name: bond0
  type: bond
  state: up
  accept-all-mac-addresses: false
  ethtool:
    feature:
        [../..]
  ipv4:
    enabled: false
    address: []
    dhcp: false
  ipv6:
    enabled: false
    address: []
    autoconf: false
    dhcp: false
  link-aggregation:
    mode: active-backup
    options:
      all_slaves_active: dropped
      arp_all_targets: any
      arp_interval: 0
      arp_validate: none
      downdelay: 0
      fail_over_mac: none
      miimon: 100
      num_grat_arp: 1
      num_unsol_na: 1
      primary: eno1
      primary_reselect: always
      resend_igmp: 1
      updelay: 0
      use_carrier: true
    port:
    - eno1
    - eno2
  lldp:
    enabled: false
  mac-address: X
  mtu: 1500

The state for eno1 and eno2 is "up".


The hypervisor which has the problem seems to be perfectly functional, ovirt-engine does not raise any problem.

This really sounds like something that a simple call to "Refresh Capabilities" could fix.

I did it several times. Everything is fetched (I checked in the logs), but the states are still down for all interfaces... If I do a fresh install in 4.4.4, the states shown by rhevm are OK, if I reinstall in 4.4.10 the WebUI Hosts/<node>/Network Interfaces is KO.
 
 
That's really strange, I would suggest removing the host completely from the engine if possible and then adding it again. That should also remove the host from DB and clear the references.

Is it only one host that's affected or multiple?

Best regards,
Ales



--

Ales Musil

Senior Software Engineer - RHV Network

Red Hat EMEA

amusil@redhat.com    IM: amusil