fresh ovirt node 4.4.6 fail on firewalld both host and engine deployment
by Charles Kozler
Hello -
Deployed fresh ovirt node 4.4.6 and the only thing I did to the system was
configure the NIC with nmtui
During the gluster install the install errored out with
gluster-deployment-1620832547044.log:failed: [n2] (item=5900/tcp) =>
{"ansible_loop_var": "item", "changed": false, "item": "5900/tcp", "msg":
"ERROR: Exception caught: org.fedoraproject.FirewallD1.Exception:
ALREADY_ENABLED: '5900:tcp' already in 'public' Permanent and
Non-Permanent(immediate) operation"}
The fix here was easy - I just deleted the port it was complaining about
with firewall-cmd and restarted the installation and it was all fine
During the hosted engine deployment when the VM is being deployed it dies
here
[ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Open a port on firewalld]
[ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": "ERROR:
Exception caught: org.fedoraproject.FirewallD1.Exception: ALREADY_ENABLED:
'6900:tcp' already in 'public' Non-permanent operation"}
Now the issue here is that I do not have access to the engine VM as it is
in a bit of a transient state since when it fails the current image that is
open is discarded when the ansible playbook is kicked off again
I cannot find any BZ on this and google is turning up nothing. I don't
think firewalld failing due to the firewall rule already existing should be
a reason to exit the installation
The interesting part is that this only fails on certain ports. i.e when I
reran the gluster wizard after 5900 failed, the other ports are presumably
still added to the firewall, and the installation completes
Suggestions?
--
*Notice to Recipient*: https://www.fixflyer.com/disclaimer
<https://www.fixflyer.com/disclaimer>
3 years, 7 months
Update of plain CentOS hosts very slow
by Gianluca Cecchi
Hello,
I have a 4.4.5 environment that I'm upgrading to 4.4.6.
I'm upgrading plain CentOS hosts from the GUI.
They are in 4.4.5, so in particular CentOS 8.3 and as part of the upgrade
they have to be put to 8.4.
In the past I used "yum update" on the host but now it seems it is not the
correct way.
But the ansible part related to package updates seems to be very slow.
It gives the impression that it is doing it one by one and not as a whole
when you run "yum update"
Now it is about 30 minutes that the update is going on and my internet
speed is for sure very high.
In messages of host I see every single line suche this ones:
Jun 8 11:09:30 ov300 python3[3031815]: ansible-dnf Invoked with
name=['rsyslog-relp.x86_64'] state=latest lock_timeout=300
conf_file=/tmp/yum.conf allow_downgrade=False autoremove=False bugfix=False
disable_gpg_check=False disable_plugin=[] disablerepo=[]
download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/
install_repoquery=True install_weak_deps=True security=False
skip_broken=False update_cache=False update_only=False validate_certs=True
disable_excludes=None download_dir=None list=None releasever=None
Jun 8 11:09:32 ov300 python3[3031828]: ansible-dnf Invoked with
name=['runc.x86_64'] state=latest lock_timeout=300 conf_file=/tmp/yum.conf
allow_downgrade=False autoremove=False bugfix=False disable_gpg_check=False
disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[]
enablerepo=[] exclude=[] installroot=/ install_repoquery=True
install_weak_deps=True security=False skip_broken=False update_cache=False
update_only=False validate_certs=True disable_excludes=None
download_dir=None list=None releasever=None
Any clarification?
Thanks,
Gianluca
3 years, 7 months
Rootless Podman container not displaying in oVirt Manager
by David White
I deployed a rootless Podman container on a RHEL 8 guest on Saturday (3 days ago).
At the time, I remember seeing some selinux AVC "denied" messages related to qemu-guest-agent and podman, but I didn't have time to look into it further, but made a mental note to come back to it, because it really smelled like a bug to me.
So, I came back to it this afternoon, and now I see nothing when I look for `ausearch -m AVC`
I restarted the `quemu-guest-agent` service with systemctl, and ran `ausearch -m AVC` again, hoping to see some results, but I still don't.
I really wish that I had at least copied the AVC message I saw on Saturday for later investigation, but I fully expected to be able to find that information again today.
Regardless, I have a rootless container running on the guest VM.
When I login to the oVirt Manager and navigate to the VM -> Containers, I don't see anything listed.
On Saturday, I thought this was a bug with selinux and qemu-guest-agent.
But now, I have no idea.
Any thoughts?
Sent with ProtonMail Secure Email.
3 years, 7 months
Booting VMs from RHEL ISOs fail
by David White
Ever since I deployed oVirt a couple months ago, I've been unable to boot any VMs from a RHEL ISO.
Ubuntu works fine, as does CentOS.
I've tried multiple RHEL 8 ISOs on multiple VMs.
I've destroyed and re-uploaded the ISOs, and I've also destroyed and re-created the VMs.
Every time I try to boot a VM to a RHEL 8 ISO, the console just tells me that "No boot device" was found.
Can anyone think of any reason why other ISOs would work, when RHEL ISOs do not work? How can I troubleshoot this further?
I really need to get a server up and running with Podman.
Sent with ProtonMail Secure Email.
3 years, 7 months
4.3 engine cert
by KSNull Zero
Hello!
oVirt 4.3 engine.cer is about to expire.
What is the proper way to renew it, so there is no impact on the running hosts and workloads ?
Thank you.
3 years, 7 months
Ovirt node 4.4.5 failure to upgrade to 4.4.6
by Guillaume Pavese
Maybe my problem is in part linked to an issue seen by Jayme earlier, but
then the resolution that worked for him did not succeed for me :
I first upgraded my Self Hosted Engine from 4.4.5 to 4.4.6 and then
upgraded it to Centos-Stream and rebooted
Then I tried to upgrade the cluster (3 ovirt-nodes on 4.4.5) but it failed
at the first host.
They are all ovir-node hosts, originally first installed in 4.4.5
In Host Event Logs I saw :
...
Update of host ps-inf-prd-kvm-fr-510.hostics.fr.
Upgrade packages
Update of host ps-inf-prd-kvm-fr-510.hostics.fr.
Check if image was updated.
Update of host ps-inf-prd-kvm-fr-510.hostics.fr.
Check if image was updated.
Update of host ps-inf-prd-kvm-fr-510.hostics.fr.
Check if image-updated file exists.
Failed to upgrade Host ps-inf-prd-kvm-fr-510.hostics.fr (User:
gpav(a)hostics.fr).
ovirt-node-ng-image-update-4.4.6.3-1.el8.noarch was installed according to
yum,
I tried reinstalling it but got errors: "Error in POSTIN scriptlet" :
Downloading Packages:
[SKIPPED] ovirt-node-ng-image-update-4.4.6.3-1.el8.noarch.rpm: Already
downloaded
...
Running scriptlet: ovirt-node-ng-image-update-4.4.6.3-1.el8.noarch
Reinstalling : ovirt-node-ng-image-update-4.4.6.3-1.el8.noarch
Running scriptlet: ovirt-node-ng-image-update-4.4.6.3-1.el8.noarch
warning: %post(ovirt-node-ng-image-update-4.4.6.3-1.el8.noarch) scriptlet
failed, exit status 1
Error in POSTIN scriptlet in rpm package ovirt-node-ng-image-update
---
Reinstalled:
ovirt-node-ng-image-update-4.4.6.3-1.el8.noarch
nodectl still showed it was on 4.4.5 :
[root@ps-inf-prd-kvm-fr-510 ~]# nodectl info
bootloader:
default: ovirt-node-ng-4.4.5.1-0.20210323.0 (4.18.0-240.15.1.el8_3.x86_64)
...
current_layer: ovirt-node-ng-4.4.5.1-0.20210323.0+1
I tried to upgrade the Host again from oVirt and this time there was no
error, and the host rebooted.
However, it did not pass active after rebooting and nodectl still shows
that it's 4.4.5 installed. Similar symptoms as OP
So I removed ovirt-node-ng-image-update, then reinstalled it and got no
error this time.
nodectl info seemed to show that it was installed :
[root@ps-inf-prd-kvm-fr-510 yum.repos.d]# nodectl info
bootloader:
default: ovirt-node-ng-4.4.6.3-0.20210518.0 (4.18.0-301.1.el8.x86_64)
...
current_layer: ovirt-node-ng-4.4.5.1-0.20210323.0+1
However, after reboot the Host was still shown as "unresponsive"
After Marking it as "Manually rebooted", passing it in maintenance mode and
trying to activate it, the Host was automatically fenced. And still
unresponsive after this new reboot.
I passed it in maintenance mode again, And tried to reinstall it with
"Deploy Hosted Engine" selected
However if failed : "Task Stop services failed to execute."
In
/var/log/ovirt-engine/host-deploy/ovirt-host-deploy-ansible-20210602082519-ps-inf-prd-kvm-fr-510.hostics.fr-0565d681-9406-4fa7-a444-7ee34804579c.log
:
"msg" : "Unable to stop service vdsmd.service: Job for vdsmd.service
canceled.\n", "failed" : true,
"msg" : "Unable to stop service supervdsmd.service: Job for
supervdsmd.service canceled.\n", failed" : true,
"stderr" : "Error: ServiceOperationError: _systemctlStop failed\nb'Job for
vdsmd.service canceled.\\n' ",
"stderr_lines" : [ "Error: ServiceOperationError: _systemctlStop failed",
"b'Job for vdsmd.service canceled.\\n' " ],
If I try on the Host I get :
[root@ps-inf-prd-kvm-fr-510 ~]# systemctl stop vdsmd
Job for vdsmd.service canceled.
[root@ps-inf-prd-kvm-fr-510 ~]# systemctl status vdsmd
● vdsmd.service - Virtual Desktop Server Manager
Loaded: loaded (/usr/lib/systemd/system/vdsmd.service; enabled; vendor
preset: disabled)
Active: deactivating (stop-sigterm) since Wed 2021-06-02 08:49:21 CEST;
7s ago
Process: 54037 ExecStartPre=/usr/libexec/vdsm/vdsmd_init_common.sh
--pre-start (code=exited, status=0/SUCCESS)
...
Jun 02 08:47:34 ps-inf-prd-kvm-fr-510.hostics.fr vdsm[54100]: WARN Failed
to retrieve Hosted Engine HA info, is Hosted Engine setup finished?
...
Jun 02 08:48:31 ps-inf-prd-kvm-fr-510.hostics.fr vdsm[54100]: WARN Worker
blocked: <Worker name=jsonrpc/4 running <Task <JsonRpcTask {'jsonrpc':
'2.0', 'method': 'StoragePool.connectStorageServer', 'params': {'storage>
File:
"/usr/lib64/python3.6/threading.py", line 884, in _bootstrap
self._bootstrap_inner()
Retrying to manually stop vdsmd a second time then seems to work...
I tried rebooting again, restarting the install always fail at the the same
spot
What should I try to get this host back up?
Guillaume Pavese
Ingénieur Système et Réseau
Interactiv-Group
--
Ce message et toutes les pièces jointes (ci-après le “message”) sont
établis à l’intention exclusive de ses destinataires et sont confidentiels.
Si vous recevez ce message par erreur, merci de le détruire et d’en avertir
immédiatement l’expéditeur. Toute utilisation de ce message non conforme a
sa destination, toute diffusion ou toute publication, totale ou partielle,
est interdite, sauf autorisation expresse. L’internet ne permettant pas
d’assurer l’intégrité de ce message . Interactiv-group (et ses filiales)
décline(nt) toute responsabilité au titre de ce message, dans l’hypothèse
ou il aurait été modifié. IT, ES, UK.
<https://interactiv-group.com/disclaimer.html>
3 years, 7 months
ovirt 4 live migration problem
by david
engine 4.4.4.7-1.el8
can someone explain to me what is the problem, when i try to migrate some
vms
to another host in the cluster the migration status reaches the percent of
99 and this error message appears:
+-------------------------+
Failed to migrate VM scom1-a66 to Host kvm3
No available host was found to migrate VM scom1-a66 to.
Trying to migrate to another Host.
+-------------------------+
i have attached engine.log where the last problematic correlation is
d448f6ad-c549-4f72-b3da-bed927f32b23
i have two node cluster
the servers(kvm4 and kvm3) in the cluster have a different hardware
configuration and kvm version
also attached virsh-capabilities and vdsm.log from them
the strangest thing is that in the logs there is no reason at all why the
migration failed. at least i didn't find anything
kvm4
====
OS Version: RHEL - 8.3 - 1.2011.el8
Kernel Version: 4.18.0 - 240.1.1.el8_3.x86_64
KVM Version: 5.1.0 - 14.el8.1
LIBVIRT Version: libvirt-6.6.0-7.1.el8
VDSM Version: vdsm-4.40.40-1.el8
SPICE Version: 0.14.3 - 3.el8
kvm3
====
OS Version: RHEL - 8.4 - 1.2105.el8
Kernel Version: 4.18.0 - 305.3.1.el8.x86_64
KVM Version: 5.1.0 - 20.el8
LIBVIRT Version: libvirt-6.6.0-13.el8
VDSM Version: vdsm-4.40.60.7-1.el8
SPICE Version: 0.14.3 - 4.el8
3 years, 7 months
I am installing ovirt engine 4.3.10
by ken@everheartpartners.com
I am getting this error message when I install it on CentOS 7.9 when running the hosted engine setup.
[ INFO ] TASK [ovirt.hosted_engine_setup : Validate selected bridge interface if management bridge does not exists]
[ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": "The selected network interface is not valid"}
I have two interfaces
enp6s0
enp11s0
Enp11s0 is the public network
enp6s0 is the storage network to the netapp.
Any idea how to resolve this?
3 years, 7 months
Fedora CoreOS
by lejeczek
Hi guys.
From what I gather there is no oVirt for Fedora CoreOS but
I should ask here at the source - is it there oVirt for that
OS and if there is not as of now, are the any plans or
discussion to make that reality?
many thanks, L.
3 years, 7 months