4.4.9 -> 4.4.10 Cannot start or migrate any VM (hotpluggable cpus requested exceeds the maximum cpus supported by KVM)
by Jillian Morgan
After upgrading the engine from 4.4.9 to 4.4.10, and then upgrading one
host, any attempt to migrate a VM to that host or start a VM on that host
results in the following error:
Number of hotpluggable cpus requested (16) exceeds the maximum cpus
supported by KVM (8)
While the version of qemu is the same across hosts, (
qemu-kvm-6.0.0-33.el8s.x86_64), I traced the difference to the upgraded
kernel on the new host. I have always run elrepo's kernel-ml on these hosts
to support bcache which RHEL's kernel doesn't support. The working hosts
still run kernel-ml-5.15.12. The upgraded host ran kernel-ml-5.17.0.
In case anyone else runs kernel-ml, have you run into this issue?
Does anyone know why KVM's KVM_CAP_MAX_VCPUS value is lowered on the new
kernel?
Does anyone know how to query the KVM capabilities from userspace without
writing a program leveraging kvm_ioctl()'s?
Related to this, it seems that ovirt and/or libvirtd always runs qmu-kvm
with an -smp argument of "maxcpus=16". This causes qemu's built-in check to
fail on the new kernel which is supporting max_vpus of 8.
Why does ovirt always request maxcpus=16?
And yes, before you say it, I know you're going to say that running
kernel-ml isn't supported.
--
Jillian Morgan (she/her) 🏳️⚧️
Systems & Networking Specialist
Primordial Software Group & I.T. Consultancy
https://www.primordial.ca
1 year, 10 months
4.5.2 Create Additional Gluster Logical Volumes fails
by simon@justconnect.ie
Hi,
In 4.4 adding additional gluster volumes was a simple ansible task (or via cockpit).
With 4.5.2 I tried to add new volumes but the logic has changed/broken. Here's the error I am getting:
TASK [gluster.infra/roles/backend_setup : Create volume groups] ********************************************************************************************************************************
failed: [bdtovirthcidmz02-strg.mydomain.com] (item={'key': 'gluster_vg_sda', 'value': [{'vgname': 'gluster_vg_sda', 'pvname': '/dev/sda'}]}) => {"ansible_loop_var": "item", "changed": true, "cmd": ["vgcreate", "--dataalignment", "2048K", "-s", "2048K", "gluster_vg_sda", "/dev/sda"], "delta": "0:00:00.010442", "end": "2022-11-10 13:11:16.717772", "item": {"key": "gluster_vg_sda", "value": [{"pvname": "/dev/sda", "vgname": "gluster_vg_sda"}]}, "msg": "non-zero return code", "rc": 3, "start": "2022-11-10 13:11:16.707330", "stderr": " Configuration setting \"filter\" invalid. It's not part of any section.\n /dev/gluster_vg_sda: already exists in filesystem\n Run `vgcreate --help' for more information.", "stderr_lines": [" Configuration setting \"filter\" invalid. It's not part of any section.", " /dev/gluster_vg_sda: already exists in filesystem", " Run `vgcreate --help' for more information."], "stdout": "", "stdout_lines": []}
failed: [bdtovirthcidmz03-strg.mydomain.com] (item={'key': 'gluster_vg_sda', 'value': [{'vgname': 'gluster_vg_sda', 'pvname': '/dev/sda'}]}) => {"ansible_loop_var": "item", "changed": true, "cmd": ["vgcreate", "--dataalignment", "2048K", "-s", "2048K", "gluster_vg_sda", "/dev/sda"], "delta": "0:00:00.010231", "end": "2022-11-10 13:12:35.607565", "item": {"key": "gluster_vg_sda", "value": [{"pvname": "/dev/sda", "vgname": "gluster_vg_sda"}]}, "msg": "non-zero return code", "rc": 3, "start": "2022-11-10 13:12:35.597334", "stderr": " Configuration setting \"filter\" invalid. It's not part of any section.\n /dev/gluster_vg_sda: already exists in filesystem\n Run `vgcreate --help' for more information.", "stderr_lines": [" Configuration setting \"filter\" invalid. It's not part of any section.", " /dev/gluster_vg_sda: already exists in filesystem", " Run `vgcreate --help' for more information."], "stdout": "", "stdout_lines": []}
failed: [bdtovirthcidmz01-strg.mydomain.com] (item={'key': 'gluster_vg_sda', 'value': [{'vgname': 'gluster_vg_sda', 'pvname': '/dev/sda'}]}) => {"ansible_loop_var": "item", "changed": true, "cmd": ["vgcreate", "--dataalignment", "2048K", "-s", "2048K", "gluster_vg_sda", "/dev/sda"], "delta": "0:00:00.011282", "end": "2022-11-10 13:13:24.336233", "item": {"key": "gluster_vg_sda", "value": [{"pvname": "/dev/sda", "vgname": "gluster_vg_sda"}]}, "msg": "non-zero return code", "rc": 3, "start": "2022-11-10 13:13:24.324951", "stderr": " Configuration setting \"filter\" invalid. It's not part of any section.\n /dev/gluster_vg_sda: already exists in filesystem\n Run `vgcreate --help' for more information.", "stderr_lines": [" Configuration setting \"filter\" invalid. It's not part of any section.", " /dev/gluster_vg_sda: already exists in filesystem", " Run `vgcreate --help' for more information."], "stdout": "", "stdout_lines": []}
The vg was created as part of the initial ansible build with logical volumes being added when required.
Any assistance would be greatly appreciated.
Kind regards
Simon
1 year, 10 months
Out-of-sync networks can only be detached
by Sakhi Hadebe
Hi,
I have a 3-node oVirt cluster. I have configured 2 logical networks:
ovirtmgmt and public. Public logical network is attached in only 2 nodes
and failing to attach on the 3rd node with the below error
Invalid operation, out-of-sync network 'public' can only be detached.
Please have been stuck on this for almost the whole day now. How do I fix
this error?
--
Regards,
Sakhi Hadebe
1 year, 11 months
Re: Self-hosted engine 4.5.0 deployment fails
by John
I would run the deploy again, wait until the engine is up and then from
the server you are deploying the engine on
# virsh list
obtain the virtual machine number from the above command, in the example
below we assume the number is 1
# virsh console 1
login as root
wait for all packages to finish updating:
tail -f /var/log/dnf.rpm.log
then:
# dnf downgrade postgresql-jdbc
once that's done the deployment of the engine should get past the
[ INFO ] TASK [ovirt.ovirt.engine_setup : Check if Engine health page is
up]
step.
On 03/05/2022 17:03, Mohamed Roushdy wrote:
>
> Hello,
>
> I’m deploying on Ovirt nodes v4.5.0.1, but it fails to deploy the
> self-hosted engine with the following errors in the web installer:
>
> [ INFO ] skipping: [localhost]
> [ INFO ] TASK [ovirt.ovirt.engine_setup : Check if Engine health page
> is up]
> [ ERROR ] fatal: [localhost -> 192.168.222.197]: FAILED! =>
> {"attempts": 30, "changed": false, "connection": "close",
> "content_encoding": "identity", "content_length": "86",
> "content_type": "text/html; charset=UTF-8", "date": "Tue, 03 May 2022
> 15:57:20 GMT", "elapsed": 0, "msg": "Status code was 500 and not
> [200]: HTTP Error 500: Internal Server Error", "redirected": false,
> "server": "Apache/2.4.37 (centos) OpenSSL/1.1.1k
> mod_auth_gssapi/1.6.1", "status": 500, "url":
> http://localhost/ovirt-engine/services/health}
> [ INFO ] TASK [ovirt.ovirt.engine_setup : Clean temporary files]
> [ INFO ] changed: [localhost -> 192.168.222.197]
> [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Sync on engine machine]
> [ INFO ] changed: [localhost -> 192.168.222.197]
> [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Set destination
> directory path]
> [ INFO ] ok: [localhost -> localhost]
> [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Create destination
> directory]
> [ INFO ] changed: [localhost -> localhost]
> [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : include_tasks]
> [ INFO ] ok: [localhost]
> [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Find the local
> appliance image]
> [ INFO ] ok: [localhost -> localhost]
> [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Set local_vm_disk_path]
> [ INFO ] ok: [localhost -> localhost]
> [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Give the vm time to
> flush dirty buffers]
> [ INFO ] ok: [localhost -> localhost]
> [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Copy engine logs]
> [ INFO ] changed: [localhost]
> [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Change ownership of
> copied engine logs]
> [ INFO ] changed: [localhost -> localhost]
> [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Notify the user about
> a failure]
> [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg":
> "There was a failure deploying the engine on the local engine VM. The
> system may not be provisioned according to the playbook results:
> please check the logs for the issue, fix accordingly or re-deploy from
> scratch.\n"}
>
> Any idea if this is a bug or something? I tried to deploy for eight
> times with different configuration
>
> Thank y9ou,
>
>
> _______________________________________________
> Users mailing list --users(a)ovirt.org
> To unsubscribe send an email tousers-leave(a)ovirt.org
> Privacy Statement:https://www.ovirt.org/privacy-policy.html
> oVirt Code of Conduct:https://www.ovirt.org/community/about/community-guidelines/
> List Archives:https://lists.ovirt.org/archives/list/users@ovirt.org/message/XL...
2 years
Changing the Cluster Compatibility Version in hyperconverged environment
by lexxthefox@gmail.com
Hi!
I'm sorry if it is a noob question but it is very important for me.
We are researching oVirt in our company and hope to use existing environment where there's no SAN/NAS solution. Just a bunch of servers. We were able to set up a hyperconverged installation with oVirt 4.4 a while ago and now I'm resuming research and trying to upgrade it to 4.5. I was able to upgrade HE and Hosts. What's left is to upgrade Cluster Compatibility Version from 4.5 to 4.7. I don't seem to find a way to do it as when I try to Edit cluster I get this error: "Cluster Compatibility Version".
Upgrade Guide on https://www.ovirt.org/documentation/upgrade_guide/ does not have entries for hyperconverged set up. So I don't know where to look for information on this topic.
Is it possible to achieve my goal at all? Can "frozen" Cluster Compatibility Version cause problems in the long run?
I can rebuild the whole setup now when it is still being reviewed and planed. But if we start using it actively it won't be possible. I'm afraid that this could be a show-stopper in the long run.
Thanks in advance!
2 years
oVirt + TrueNAS: Unable to create iSCSI domain - I am missing something obvious
by David Johnson
Good morning folks, and thank you in advance.
I am working on migrating my oVirt backing store from NFS to iSCSI.
*oVirt Environment:*
oVirt Open Virtualization Manager
Software Version:4.4.4.7-1.el8
*TrueNAS environment:*
FreeBSD truenas.local 12.2-RELEASE-p11 75566f060d4(HEAD) TRUENAS amd64
The iSCSI share is on a TrueNAS server, exposed to user VDSM and group 36.
oVirt sees the targeted share, but is unable to make use of it.
The latest issue is "Error while executing action New SAN Storage Domain:
Volume Group block size error, please check your Volume Group
configuration, Supported block size is 512 bytes."
As near as I can tell, oVirt does not support any block size other than 512
bytes, while TrueNAS's smallest OOB block size is 4k.
I know that oVirt on TrueNAS is a common configuration, so I expect I am
missing something really obvious here, probably a TrueNAS configuration
needed to make TrueNAS work with 512 byte blocks.
Any advice would be helpful.
*David Johnson*
2 years
update path from 4.4.4 to 4.5.4
by marek
hi,
i have standalone ovirt-engine 4.4.4 (centos8 stream)
few hosts (mix of old centos 8.2, centos 8 stream, rocky linux)
can you confirm if this upgrade path is ok?
first round
- update engine to 4.4.10
- update OS on engine to last centos 8 stream (can i upgrade to some
point in time recommended for 4.4.10?)
- update hosts to ovirt 4.4.10
- update OS on hosts to last centos 8 stream and rocky linux 8.6
special case in first round - centos 8.2
- migrate OS on hosts from centos 8.2 to rocky linux 8.6
- update hosts to ovirt 4.4.10
second round
- update engine to 4.5.4
- update hosts to ovirt 4.5.4
- update OS on hosts - centos 8 stream - is ALREADY on last update from
first round
- update OS on hosts - rocky linux - from 8.6 to 8.7
Marek
2 years
noVNC console error : Something went wrong, connection is closed
by dvx.mellin@gmail.com
Hello everyone,
Context :
oVirt 4.4.9.3-1.el8
glusterfs 8.6
Self-hosted engine
Problem:
Since I update my certificates (with engine-setup) because of the warning that my certs will expire soon, I can't open the noVNC console anymore.
In the logs on engine, I have these :
in /var/log/messages :
ovsdb-server[510110]: ovs|04628|jsonrpc|WARN|ssl:[::ffff:<ip_node_1>]:51214: receive error: Protocol error
ovsdb-server[510110]: ovs|04629|reconnect|WARN|ssl:[::ffff:<ip_node_1>]:51214: connection dropped (Protocol error)
journal[516217]: 2022-12-21 11:31:27,800+0100 ovirt-websocket-proxy: INFO msg:871 handler exception: [SSL: SSLV3_ALERT_CERTIFICATE_UNKNOWN] sslv3 alert certificate unknown (_ssl.c:897)
ovirt-websocket-proxy.py[509812]: ovirt-websocket-proxy[516217] INFO msg:871 handler exception: [SSL: SSLV3_ALERT_CERTIFICATE_UNKNOWN] sslv3 alert certificate unknown (_ssl.c:897)
ovsdb-server[510110]: ovs|04632|stream_ssl|WARN|SSL_accept: error:1417C086:SSL routines:tls_process_client_certificate:certificate verify failed
in /var/log/openvswitch/ovsdb-server-sb.log :
2022-12-21T10:31:22.540Z|04626|stream_ssl|WARN|SSL_accept: error:1417C086:SSL routines:tls_process_client_certificate:certificate verify failed
2022-12-21T10:31:22.541Z|04627|jsonrpc|WARN|Dropped 1 log messages in last 8 seconds (most recently, 8 seconds ago) due to excessive rate
2022-12-21T10:31:22.541Z|04628|jsonrpc|WARN|ssl:[::ffff:<ip_node_1>]:51214: receive error: Protocol error
2022-12-21T10:31:22.542Z|04629|reconnect|WARN|ssl:[::ffff:<ip_node_1>]:51214: connection dropped (Protocol error)
I've tried these commands (found here : https://access.redhat.com/solutions/6877501) :
/usr/share/ovirt-engine/bin/pki-enroll-pkcs12.sh --name="ovirt-provider-ovn" --password=mypass --subject="<subject_engine>" --keep-key
/usr/share/ovirt-engine/bin/pki-enroll-pkcs12.sh --name="ovn-ndb" --password=mypass --subject="<subject_engine>" --keep-key
/usr/share/ovirt-engine/bin/pki-enroll-pkcs12.sh --name="ovn-sdb" --password=mypass --subject="<subject_engine>" --keep-key
systemctl restart ovirt-provider-ovn.service
systemctl restart ovn-northd.service
Still not work, so I've seen that some certificates was still not renewed :
/usr/share/ovirt-engine/bin/pki-enroll-pkcs12.sh --name="vmconsole-proxy-helper" --password=mypass --subject="<subject_engine>" --keep-key
/usr/share/ovirt-engine/bin/pki-enroll-pkcs12.sh --name="vmconsole-proxy-host" --password=mypass --subject="<subject_engine>" --keep-key
/usr/share/ovirt-engine/bin/pki-enroll-pkcs12.sh --name="vmconsole-proxy-user" --password=mypass --subject="<subject_engine>" --keep-key
And restart every ovirt services.
But it still does not work better.
I don't see any other unvalid certificates in /etc/pki/ovirt-engine/certs/, so I don't know which certificate is invalid for ovsdb.
Thanks for any advice.
Best regards,
Michael
2 years, 1 month