Unable to reinstall 2nd and 3rd Host after restore HE backup onto 1st host
by James Wadsworth
Hi there,
We are running ovirt 4.5.4 on 3 RHEL 8.7 host with a on self hosted engine again using RHEL 8.7. We are making some changes to our company network so we followed these instructions here to move the ovirt engine onto a new storage with different IP address. https://access.redhat.com/solutions/6529691
We used a host to redeploy the hosted engine from a backup and we managed to get the hosted engine running on the new storage with the IP address. We then tried to reinstall the 2 other hosts but the reinstallation failed with the message:
Host ovirt2.ad.tintolav.com installation failed. Task Restart services failed to execute. Please check logs for more details: /var/log/ovirt-engine/host-deploy/ovirt-host-deploy-ansible-20230323083407-ovirt2.ad.tintolav.com-d950815b-f1b9-4dcf-b609-3ff1866a6b70.log
In the ovirt-engie host deploy log we see [ "Traceback (most recent call last):", " File \"/usr/bin/vdsm-tool\", line 18, in <module>", " import vdsm.tool", "ModuleNotFoundError: No module named 'vdsm'" ]
In the vdsm log on the host there are no errors except for a couple of warnings
2023-03-23 07:58:22,335+0100 WARN (periodic/1) [throttled] MOM not available. Error: [Errno 2] No such file or directory (throttledlog:87)
2023-03-23 07:58:22,336+0100 WARN (periodic/1) [throttled] MOM not available, KSM stats will be missing. Error: (throttledlog:87)
We have tried a completely new install for ovirt2, but we have finished back at the same point with the same error.
[root@ovirt2 ~]# systemctl status vdsmd
● vdsmd.service - Virtual Desktop Server Manager
Loaded: loaded (/usr/lib/systemd/system/vdsmd.service; enabled; vendor preset: disabled)
Active: inactive (dead) since Thu 2023-03-23 08:35:32 CET; 56min ago
Process: 22942 ExecStart=/usr/libexec/vdsm/daemonAdapter -0 /dev/null -1 /dev/null -2 /dev/null /usr/libexec/vdsm/vdsmd (code=exited, status=0/SUCCESS)
Process: 22876 ExecStartPre=/usr/libexec/vdsm/vdsmd_init_common.sh --pre-start (code=exited, status=0/SUCCESS)
Main PID: 22942 (code=exited, status=0/SUCCESS)
Mar 23 07:58:20 ovirt2.ad.tintolav.com vdsmd_init_common.sh[22876]: vdsm: Running test_space
Mar 23 07:58:20 ovirt2.ad.tintolav.com vdsmd_init_common.sh[22876]: vdsm: Running test_lo
Mar 23 07:58:20 ovirt2.ad.tintolav.com systemd[1]: Started Virtual Desktop Server Manager.
Mar 23 07:58:22 ovirt2.ad.tintolav.com vdsm[22942]: WARN MOM not available. Error: [Errno 2] No such file or directory
Mar 23 07:58:22 ovirt2.ad.tintolav.com vdsm[22942]: WARN MOM not available, KSM stats will be missing. Error:
Mar 23 08:35:32 ovirt2.ad.tintolav.com systemd[1]: Stopping Virtual Desktop Server Manager...
Mar 23 08:35:32 ovirt2.ad.tintolav.com systemd[1]: vdsmd.service: Succeeded.
Mar 23 08:35:32 ovirt2.ad.tintolav.com systemd[1]: Stopped Virtual Desktop Server Manager.
Mar 23 08:35:33 ovirt2.ad.tintolav.com systemd[1]: Dependency failed for Virtual Desktop Server Manager.
Mar 23 08:35:33 ovirt2.ad.tintolav.com systemd[1]: vdsmd.service: Job vdsmd.service/start failed with result 'dependency'.
Mar 23 08:35:33 ovirt2.ad.tintolav.com systemd[1]: vdsmd.service: Job vdsmd.service/start failed with result 'dependency'.
[root@ovirt2 ~]# systemctl start vdsmd
A dependency job for vdsmd.service failed. See 'journalctl -xe' for details.
[root@ovirt2 ~]# journalctl -xe
Mar 23 09:34:35 ovirt2.ad.tintolav.com systemd[1]: Stopped Auxiliary vdsm service for running helper functions as root.
-- Subject: Unit supervdsmd.service has finished shutting down
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit supervdsmd.service has finished shutting down.
Mar 23 09:34:35 ovirt2.ad.tintolav.com systemd[1]: Started Auxiliary vdsm service for running helper functions as root.
-- Subject: Unit supervdsmd.service has finished start-up
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit supervdsmd.service has finished starting up.
--
-- The start-up result is done.
Mar 23 09:34:35 ovirt2.ad.tintolav.com daemonAdapter[28677]: Traceback (most recent call last):
Mar 23 09:34:35 ovirt2.ad.tintolav.com daemonAdapter[28677]: File "/usr/libexec/vdsm/daemonAdapter", line 16, in <module>
Mar 23 09:34:35 ovirt2.ad.tintolav.com daemonAdapter[28677]: from vdsm.config import config
Mar 23 09:34:35 ovirt2.ad.tintolav.com daemonAdapter[28677]: ModuleNotFoundError: No module named 'vdsm'
Mar 23 09:34:35 ovirt2.ad.tintolav.com systemd[1]: supervdsmd.service: Main process exited, code=exited, status=1/FAILURE
Mar 23 09:34:35 ovirt2.ad.tintolav.com systemd[1]: supervdsmd.service: Failed with result 'exit-code'.
-- Subject: Unit failed
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit supervdsmd.service has entered the 'failed' state with result 'exit-code'.
Mar 23 09:34:35 ovirt2.ad.tintolav.com systemd[1]: supervdsmd.service: Service RestartSec=100ms expired, scheduling restart.
Mar 23 09:34:35 ovirt2.ad.tintolav.com systemd[1]: supervdsmd.service: Scheduled restart job, restart counter is at 5.
-- Subject: Automatic restarting of a unit has been scheduled
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Automatic restarting of the unit supervdsmd.service has been scheduled, as the result for
-- the configured Restart= setting for the unit.
Mar 23 09:34:35 ovirt2.ad.tintolav.com systemd[1]: Stopped Auxiliary vdsm service for running helper functions as root.
-- Subject: Unit supervdsmd.service has finished shutting down
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit supervdsmd.service has finished shutting down.
Mar 23 09:34:35 ovirt2.ad.tintolav.com systemd[1]: supervdsmd.service: Start request repeated too quickly.
Mar 23 09:34:35 ovirt2.ad.tintolav.com systemd[1]: supervdsmd.service: Failed with result 'exit-code'.
-- Subject: Unit failed
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- The unit supervdsmd.service has entered the 'failed' state with result 'exit-code'.
Mar 23 09:34:35 ovirt2.ad.tintolav.com systemd[1]: Failed to start Auxiliary vdsm service for running helper functions as root.
-- Subject: Unit supervdsmd.service has failed
-- Defined-By: systemd
-- Support: https://access.redhat.com/support
--
-- Unit supervdsmd.service has failed.
--
-- The result is failed.
It gives the same error on the 3rd host.
Is this a python dependency issue? We are currently using python3.9. Does anyone have any ideas how we get host 2 and 3 up and running again in the cluster?
Thanks James
8 hours, 44 minutes
Re: ovirt and network outages
by Gianluca Cecchi
On Mon, Mar 20, 2023 at 7:59 PM Fedele Stabile <fedele.stabile(a)fis.unical.it>
wrote:
> Ok i will prepare all to the next outage
>
If the latest outage was not long ago, you could still find some
preliminary and useful information inside /var/log/ovirt-engine/engine.log
of the engine
In general, for all oVirt related log files you can refer to this url:
https://www.ovirt.org/documentation/administration_guide/#chap-Log_Files
some of them are also rotated
Gianluca
1 day, 1 hour
How to delete/detach a disk from online VM ?
by morgan cox
Hello.
Is it possible to remove a (unmounted, unused) disk from a VM when online?
I tried via the UI and I get 'Error while executing action: Cannot remove Virtual Disk. At least one of the VMs is not down.'
I know you can with standard KVM/Libvirt/virt-manager - is there any way I can remove a disk from a VM whilst it is online in Ovirt ?
Thanks
1 day, 5 hours
I can't enable cinderlib
by fsayd@derecho.uncu.edu.ar
Hello:
I am trying to deploy a 4.5 cluster with 3 hosts that also provides Ceph
storage. I am running the engine in a standalone host (No hosted
engine).
I followed the steps to enable cinderlib
(https://access.redhat.com/solutions/5605141) but the managed domain
option doesn't show in the add storage domain dialog.
I don't know what I am doing wrong
Do I need to add a primary storage domain (iSCSI, NFS, etc) in order to
add a Ceph domain??
* engine: 4.5.4-1.el8
* hosts: Rocky Linux release 8.7
Thanks
engine: 4.5.4-1.el8
hosts: Rocky Linux release 8.7
2 days
oVirt 4.5.x User Creation with Keycloak
by simon@justconnect.ie
Can anyone point me in the direction of documentation to create new users and groups to provide specific permissions please?
Thanks
Simon...
2 days, 7 hours
HostedEngine restarts from time to time
by ziyi Liu
Version 4.5.3.2-1.el8
There are two red warnings in the /var/log/messages file
kernel
shpchp 0000:01:00.0: Slot initialization failed
kernel
shpchp 0000:01:00.0: pci_hp_register failed with error -16
2 days, 13 hours
Gluster Geo-Replication session not visible in oVirt Manager UI
by simon@justconnect.ie
I have previously created Gluster geo-replicated sessions from the command line on 4.4.5. These sessions appear in the 'Geo-Replication' Tab of the Gluster Volume on the Master.
I've just built 2 new oVirt clusters 4.5.3.2 and configured a geo-replicated session. The session doesn't appear in the WebUI but appears to be working when checked from the command line.
Any ideas anyone please?
Regards
Simon
2 days, 17 hours
4.5.2 Create Additional Gluster Logical Volumes fails
by simon@justconnect.ie
Hi,
In 4.4 adding additional gluster volumes was a simple ansible task (or via cockpit).
With 4.5.2 I tried to add new volumes but the logic has changed/broken. Here's the error I am getting:
TASK [gluster.infra/roles/backend_setup : Create volume groups] ********************************************************************************************************************************
failed: [bdtovirthcidmz02-strg.mydomain.com] (item={'key': 'gluster_vg_sda', 'value': [{'vgname': 'gluster_vg_sda', 'pvname': '/dev/sda'}]}) => {"ansible_loop_var": "item", "changed": true, "cmd": ["vgcreate", "--dataalignment", "2048K", "-s", "2048K", "gluster_vg_sda", "/dev/sda"], "delta": "0:00:00.010442", "end": "2022-11-10 13:11:16.717772", "item": {"key": "gluster_vg_sda", "value": [{"pvname": "/dev/sda", "vgname": "gluster_vg_sda"}]}, "msg": "non-zero return code", "rc": 3, "start": "2022-11-10 13:11:16.707330", "stderr": " Configuration setting \"filter\" invalid. It's not part of any section.\n /dev/gluster_vg_sda: already exists in filesystem\n Run `vgcreate --help' for more information.", "stderr_lines": [" Configuration setting \"filter\" invalid. It's not part of any section.", " /dev/gluster_vg_sda: already exists in filesystem", " Run `vgcreate --help' for more information."], "stdout": "", "stdout_lines": []}
failed: [bdtovirthcidmz03-strg.mydomain.com] (item={'key': 'gluster_vg_sda', 'value': [{'vgname': 'gluster_vg_sda', 'pvname': '/dev/sda'}]}) => {"ansible_loop_var": "item", "changed": true, "cmd": ["vgcreate", "--dataalignment", "2048K", "-s", "2048K", "gluster_vg_sda", "/dev/sda"], "delta": "0:00:00.010231", "end": "2022-11-10 13:12:35.607565", "item": {"key": "gluster_vg_sda", "value": [{"pvname": "/dev/sda", "vgname": "gluster_vg_sda"}]}, "msg": "non-zero return code", "rc": 3, "start": "2022-11-10 13:12:35.597334", "stderr": " Configuration setting \"filter\" invalid. It's not part of any section.\n /dev/gluster_vg_sda: already exists in filesystem\n Run `vgcreate --help' for more information.", "stderr_lines": [" Configuration setting \"filter\" invalid. It's not part of any section.", " /dev/gluster_vg_sda: already exists in filesystem", " Run `vgcreate --help' for more information."], "stdout": "", "stdout_lines": []}
failed: [bdtovirthcidmz01-strg.mydomain.com] (item={'key': 'gluster_vg_sda', 'value': [{'vgname': 'gluster_vg_sda', 'pvname': '/dev/sda'}]}) => {"ansible_loop_var": "item", "changed": true, "cmd": ["vgcreate", "--dataalignment", "2048K", "-s", "2048K", "gluster_vg_sda", "/dev/sda"], "delta": "0:00:00.011282", "end": "2022-11-10 13:13:24.336233", "item": {"key": "gluster_vg_sda", "value": [{"pvname": "/dev/sda", "vgname": "gluster_vg_sda"}]}, "msg": "non-zero return code", "rc": 3, "start": "2022-11-10 13:13:24.324951", "stderr": " Configuration setting \"filter\" invalid. It's not part of any section.\n /dev/gluster_vg_sda: already exists in filesystem\n Run `vgcreate --help' for more information.", "stderr_lines": [" Configuration setting \"filter\" invalid. It's not part of any section.", " /dev/gluster_vg_sda: already exists in filesystem", " Run `vgcreate --help' for more information."], "stdout": "", "stdout_lines": []}
The vg was created as part of the initial ansible build with logical volumes being added when required.
Any assistance would be greatly appreciated.
Kind regards
Simon
2 days, 17 hours
oVirt On Rocky 8.x - Upgrade To Rocky 9.1
by Matthew J Black
Hi All,
Sorry if this was mentioned previously (I obviously missed it if it was) but can we upgrade an oVirt (latest version) Host/Cluster and/or the oVirt Engine VM from Rocky Linux (RHEL) v8.6/8.7 to v9.1 (yet), and if so, what is / where can I find the procedure to do this - ie is there anything "special" that needs to be done because of oVirt, or can we just do a "simple" v8.x +> v9.1 upgrade?
Thanks in advance
Cheers
Dulux-Oz
2 days, 21 hours
Re: ovirt and network outages
by Gianluca Cecchi
Il Lun 20 Mar 2023, 17:32 Fedele Stabile <fedele.stabile(a)fis.unical.it> ha
scritto:
> Thank you Gianluca for your answer.
> Population of /etc/hosts as a backup with host and ovirt engine ip:
> this has already been done.
>
> Fedele
>
> ______________
>
Fedele, I think you should define better what "misaligned" means in your
case, give some details on versions in play, architecture (self hosted vs
standalone engine) and provide some logs from the engine and hosts
(preferably through a link to a file sharing service if big), so that you
can get better help.
Gianluca
>
2 days, 22 hours