Please, Please Help - New oVirt Install/Deployment Failing - "Host is not up..."

Hi Everyone, Could someone please help me - I've been trying to do an install of oVirt for *weeks* (including false starts and self-inflicted wounds/errors) and it is still not working. My setup: - oVirt v4.5.3 - A brand new fresh vanilla install of RockyLinux 8.6 - all working AOK - 2*NICs in a bond (802.3ad) with a couple of sub-Interfaces/VLANs - all working AOK - All relevant IPv4 Address in DNS with Reverse Lookups - all working AOK - All relevant IPv4 Address in "/etc/hosts" file - all working AOK - IPv6 (using "method=auto" in the interface config file) enabled on the relevant sub-Interface/VLAN - I'm not using IPv6 on the network, only IPv4, but I'm trying to cover all the bases. - All relevant Ports (as per the oVirt documentation) set up on the firewall - ie firewall-cmd --add-service={{ libvirt-tls | ovirt-imageio | ovirt-vmconsole | vdsm }} - All the relevant Repositories installed (ie RockyLinux BaseOS, AppStream, & PowerTools, and the EPEL, plus the ones from the oVirt documentation) I have followed the oVirt documentation (including the special RHEL-instructions and RockyLinux-instructions) to the letter - no deviations, no special settings, exactly as they are written. All the dnf installs, etc, went off without a hitch, including the "dnf install centos-release-ovirt45", "dnf install ovirt-engine-appliance", and "dnf install ovirt-hosted-engine-setup" - no errors anywhere. Here is the results of a "dnf repolist": - appstream Rocky Linux 8 - AppStream - baseos Rocky Linux 8 - BaseOS - centos-ceph-pacific CentOS-8-stream - Ceph Pacific - centos-gluster10 CentOS-8-stream - Gluster 10 - centos-nfv-openvswitch CentOS-8 - NFV OpenvSwitch - centos-opstools CentOS-OpsTools - collectd - centos-ovirt45 CentOS Stream 8 - oVirt 4.5 - cs8-extras CentOS Stream 8 - Extras - cs8-extras-common CentOS Stream 8 - Extras common packages - epel Extra Packages for Enterprise Linux 8 - x86_64 - epel-modular Extra Packages for Enterprise Linux Modular 8 - x86_64 - ovirt-45-centos-stream-openstack-yoga CentOS Stream 8 - oVirt 4.5 - OpenStack Yoga Repository - ovirt-45-upstream oVirt upstream for CentOS Stream 8 - oVirt 4.5 - powertools Rocky Linux 8 - PowerTools So I kicked-off the oVirt deployment with: "hosted-engine --deploy --4 --ansible-extra-vars=he_offline_deployment=true". I used "--ansible-extra-vars=he_offline_deployment=true" because without that flag I was getting "DNF timout" issues (see my previous post `Local (Deployment) VM Can't Reach "centos-ceph-pacific" Repo`). I answer the defaults to all of questions the script asked, or entered the deployment-relevant answers where appropriate. In doing this I double-checked every answer before hitting <Enter>. Everything progressed smoothly until the deployment reached the "Wait for the host to be up" task... which then hung for more than 30 minutes before failing. From the ovirt-hosted-engine-setup... log file: - 2022-10-20 17:54:26,285+1100 ERROR otopi.ovirt_hosted_engine_setup.ansible_utils ansible_utils._process_output:113 fatal: [localhost]: FAILED! => {"changed": false, "msg": "Host is not up, please check logs, perhaps also on the engine machine"} I checked the following log files and found all of the relevant ERROR lines, then checked several 10s of proceeding and succeeding lines trying to determine what was going wrong, but I could not determine anything. - ovirt-hosted-engine-setup... - ovirt-hosted-engine-setup-ansible-bootstrap_local_vm... - ovirt-hosted-engine-setup-ansible-final_clean... - not really relevant, I believe I can include the log files (or the relevant parts of the log files) if people want - but that are very large: several 100 kilobytes each. I also googled "oVirt Host is not up" and found several entries, but after reading them all the most relevant seems to be a thread from these mailing list: `Install of RHV 4.4 failing - "Host is not up, please check logs, perhaps also on the engine machine"` - but this seems to be talking about an upgrade and I didn't gleam anything useful from it - I could, of course, be wrong about that. So my questions are: - Where else should I be looking (ie other log files, etc, and possible where to find them)? - Does anyone have any idea why this isn't working? - Does anyone have a work-around (including a completely manual process to get things working - I don't mind working in the CLI with virsh, etc)? - What am I doing wrong? Please, I'm really stumped with this, and I really do need help. Cheers Dulux-Oz

On Thu, Oct 20, 2022 at 1:59 PM Matthew J Black <matthew@peregrineit.net> wrote:
Hi Everyone,
Could someone please help me - I've been trying to do an install of oVirt for *weeks* (including false starts and self-inflicted wounds/errors) and it is still not working.
My setup:
- oVirt v4.5.3 - A brand new fresh vanilla install of RockyLinux 8.6 - all working AOK - 2*NICs in a bond (802.3ad) with a couple of sub-Interfaces/VLANs - all working AOK - All relevant IPv4 Address in DNS with Reverse Lookups - all working AOK - All relevant IPv4 Address in "/etc/hosts" file - all working AOK - IPv6 (using "method=auto" in the interface config file) enabled on the relevant sub-Interface/VLAN - I'm not using IPv6 on the network, only IPv4, but I'm trying to cover all the bases. - All relevant Ports (as per the oVirt documentation) set up on the firewall - ie firewall-cmd --add-service={{ libvirt-tls | ovirt-imageio | ovirt-vmconsole | vdsm }} - All the relevant Repositories installed (ie RockyLinux BaseOS, AppStream, & PowerTools, and the EPEL, plus the ones from the oVirt documentation)
I have followed the oVirt documentation (including the special RHEL-instructions and RockyLinux-instructions) to the letter - no deviations, no special settings, exactly as they are written.
All the dnf installs, etc, went off without a hitch, including the "dnf install centos-release-ovirt45", "dnf install ovirt-engine-appliance", and "dnf install ovirt-hosted-engine-setup" - no errors anywhere.
Here is the results of a "dnf repolist":
- appstream Rocky Linux 8 - AppStream - baseos Rocky Linux 8 - BaseOS - centos-ceph-pacific CentOS-8-stream - Ceph Pacific - centos-gluster10 CentOS-8-stream - Gluster 10 - centos-nfv-openvswitch CentOS-8 - NFV OpenvSwitch - centos-opstools CentOS-OpsTools - collectd - centos-ovirt45 CentOS Stream 8 - oVirt 4.5 - cs8-extras CentOS Stream 8 - Extras - cs8-extras-common CentOS Stream 8 - Extras common packages - epel Extra Packages for Enterprise Linux 8 - x86_64 - epel-modular Extra Packages for Enterprise Linux Modular 8 - x86_64 - ovirt-45-centos-stream-openstack-yoga CentOS Stream 8 - oVirt 4.5 - OpenStack Yoga Repository - ovirt-45-upstream oVirt upstream for CentOS Stream 8 - oVirt 4.5 - powertools Rocky Linux 8 - PowerTools
So I kicked-off the oVirt deployment with: "hosted-engine --deploy --4 --ansible-extra-vars=he_offline_deployment=true".
I used "--ansible-extra-vars=he_offline_deployment=true" because without that flag I was getting "DNF timout" issues (see my previous post `Local (Deployment) VM Can't Reach "centos-ceph-pacific" Repo`).
I answer the defaults to all of questions the script asked, or entered the deployment-relevant answers where appropriate. In doing this I double-checked every answer before hitting <Enter>. Everything progressed smoothly until the deployment reached the "Wait for the host to be up" task... which then hung for more than 30 minutes before failing.
From the ovirt-hosted-engine-setup... log file:
- 2022-10-20 17:54:26,285+1100 ERROR otopi.ovirt_hosted_engine_setup.ansible_utils ansible_utils._process_output:113 fatal: [localhost]: FAILED! => {"changed": false, "msg": "Host is not up, please check logs, perhaps also on the engine machine"}
I checked the following log files and found all of the relevant ERROR lines, then checked several 10s of proceeding and succeeding lines trying to determine what was going wrong, but I could not determine anything.
- ovirt-hosted-engine-setup... - ovirt-hosted-engine-setup-ansible-bootstrap_local_vm... - ovirt-hosted-engine-setup-ansible-final_clean... - not really relevant, I believe
I can include the log files (or the relevant parts of the log files) if people want - but that are very large: several 100 kilobytes each.
You can upload to some file sharing service (dropbox, google drive, etc.) and share a link. If it's just not-too-long text you can use some pastebin (might be more or less convenient for some readers, I didn't do a poll).
I also googled "oVirt Host is not up" and found several entries, but after reading them all the most relevant seems to be a thread from these mailing list: `Install of RHV 4.4 failing - "Host is not up, please check logs, perhaps also on the engine machine"` - but this seems to be talking about an upgrade and I didn't gleam anything useful from it - I could, of course, be wrong about that.
So my questions are:
- Where else should I be looking (ie other log files, etc, and possible where to find them)?
As the error message says - check also the engine machine. Logs from it are copied to under /var/log/ovirt-hosted-engine-setup/engine-logs* by the deploy process, but you might find more by ssh to it - search local_vm_ip in the setup logs to find its local IP address. See also e.g. thread "[ovirt-users] hosted-engine-setup --deploy fail on Centos Stream 8".
- Does anyone have any idea why this isn't working?
Not without more information, sorry.
- Does anyone have a work-around (including a completely manual process to get things working - I don't mind working in the CLI with virsh, etc)?
If all you want is a workaround, I think you should already had one, no? At that point, the deploy process should not have failed, but emitted something like: You can now connect to {{ bootstrap_engine_url }} and check the status of this host and eventually remediate it, please continue only when the host is listed as 'up' Did you get this? Did you try to login? If not, there might have been some other issue - please search the log for he_pause_after_failed_add_host and host_result_up_check.
- What am I doing wrong?
Please, I'm really stumped with this, and I really do need help.
Good luck and best regards, -- Didi

OK, so I think I've identified the issue. I found these lines in the engine.log file: 2022-10-20 17:27:36,785+11 ERROR [org.ovirt.engine.core.bll.hostdeploy.AddVdsCommand] (default task-1) [4a1c2931-f754-4020-8bf3-27455338ca6d] Failed to authenticate session with host 'ovirt_node_1.mynet.local': SSH authentication to 'root@ovirt_node_1.mynet.local' failed. Please verify provided credentials. Make sure key is authorized at host 2022-10-20 17:27:36,786+11 WARN [org.ovirt.engine.core.bll.hostdeploy.AddVdsCommand] (default task-1) [4a1c2931-f754-4020-8bf3-27455338ca6d] Validation of action 'AddVds' failed for user admin@internal-authz. Reasons: VAR__ACTION__ADD,VAR__TYPE__HOST,$server ovirt_node_1.mynet.local,VDS_CANNOT_AUTHENTICATE_TO_SERVER 2022-10-20 17:27:36,822+11 ERROR [org.ovirt.engine.api.restapi.resource.AbstractBackendResource] (default task-1) [] Operation Failed: [Cannot add Host. SSH authentication failed, verify authentication parameters are correct (Username/Password, public-key etc.) You may refer to the engine.log file for further details.] So my next set of questions: - I don't recall anywhere in the deployment script being asked to supply the root login credentials for ovirt_node_1.mynet.local (the host), so what have I missed? - I don't recall being asked to supply a ssh cert for logging into ovirt_node_1.mynet.local. The only ssh cert being asked about (and its marked as "Optional") is one to log into the deployed engine *after* it is deployed (Step 16 of "5.3. Deploying the self-hosted engine using the command line"). Am I missing something here? - In the Prerequisites section of 5.3 the doco mentions "The self-hosted engine setup script requires ssh public key access using 2048-bit RSA keys from the engine virtual machine to the root account of its bare metal host." and then talks about having various settings set in the host's sshd_config file. In the host's authoized_keys file there is an RSA key (not placed by me) which I assume the deployment script placed there. Again, what am I not understanding about this? I've checked the host's sshd_config file and all of the requirements listed in section 5.3 of the doco are there (the PubkeyAcceptedKeyTypes is being picked up from the system-wide crypto policy, which are set to DEFAULT, which meets the specified 2048-bit RSA keys). I've included all the logs with the string "ERROR" in this DropBox folder, along with a couple of other relevant(?) files: https://www.dropbox.com/sh/eymwdy8hzn3sa7z/AACscSP2eaFfoiN-QzyeEVfaa?dl=0 Could someone please be kind enough to point out where things are going wrong? Cheers Dulux-Oz

Thanks for all the help Jude (including the OOB stuff). Unfortunately, this did not resolve the issue, so I'm still seeking help.

Hi, It seems that you are here: https://github.com/oVirt/ovirt-ansible-collection/blob/33d273e2e823bf0c2440a... I would say that in this case following is likely: 1) virsh list (with authentication) shows that hostedEnginelocal is running 2) SSH is successful from ovirt host due to temporary entry in /etc/hosts 3) You might see that ovirt-engine service is running on port 443 and then it is exposed as port 6900 on your ovirt host. Check out https://github.com/oVirt/ovirt-ansible-collection/blob/33d273e2e823bf0c2440a... That means you can provide he_pause_host Ansible variable with value 'true' and take a pause for troubleshooting here. So in case above is correct for your situation - then you should do port forwarding from port 6900 to port 443 on your local workstation/somewhere where you have browser and connect to ovengine GUI Then you go to 'Hosts' and see if the host is 'up'. It is unlikely that it will be up. Alternatively, there is REST API on this I think (=the one used by Ansible module in 'wait for host is up' task), but I don't know specifics. In GUI go to 'Setup host network' and make sure that network bridge/network interfaces and maybe try to assign this network (be careful: you might also connectivity to your ovirt host). Check those DHCP/Static IP assignment aspect. There could be too many assumptions above, but hopefully it gives you something to use within troubleshooting. BR, Konstantin On 20.10.22, 12:58, "Matthew J Black" <matthew@peregrineit.net> wrote: Hi Everyone, Could someone please help me - I've been trying to do an install of oVirt for *weeks* (including false starts and self-inflicted wounds/errors) and it is still not working. My setup: - oVirt v4.5.3 - A brand new fresh vanilla install of RockyLinux 8.6 - all working AOK - 2*NICs in a bond (802.3ad) with a couple of sub-Interfaces/VLANs - all working AOK - All relevant IPv4 Address in DNS with Reverse Lookups - all working AOK - All relevant IPv4 Address in "/etc/hosts" file - all working AOK - IPv6 (using "method=auto" in the interface config file) enabled on the relevant sub-Interface/VLAN - I'm not using IPv6 on the network, only IPv4, but I'm trying to cover all the bases. - All relevant Ports (as per the oVirt documentation) set up on the firewall - ie firewall-cmd --add-service={{ libvirt-tls | ovirt-imageio | ovirt-vmconsole | vdsm }} - All the relevant Repositories installed (ie RockyLinux BaseOS, AppStream, & PowerTools, and the EPEL, plus the ones from the oVirt documentation) I have followed the oVirt documentation (including the special RHEL-instructions and RockyLinux-instructions) to the letter - no deviations, no special settings, exactly as they are written. All the dnf installs, etc, went off without a hitch, including the "dnf install centos-release-ovirt45", "dnf install ovirt-engine-appliance", and "dnf install ovirt-hosted-engine-setup" - no errors anywhere. Here is the results of a "dnf repolist": - appstream Rocky Linux 8 - AppStream - baseos Rocky Linux 8 - BaseOS - centos-ceph-pacific CentOS-8-stream - Ceph Pacific - centos-gluster10 CentOS-8-stream - Gluster 10 - centos-nfv-openvswitch CentOS-8 - NFV OpenvSwitch - centos-opstools CentOS-OpsTools - collectd - centos-ovirt45 CentOS Stream 8 - oVirt 4.5 - cs8-extras CentOS Stream 8 - Extras - cs8-extras-common CentOS Stream 8 - Extras common packages - epel Extra Packages for Enterprise Linux 8 - x86_64 - epel-modular Extra Packages for Enterprise Linux Modular 8 - x86_64 - ovirt-45-centos-stream-openstack-yoga CentOS Stream 8 - oVirt 4.5 - OpenStack Yoga Repository - ovirt-45-upstream oVirt upstream for CentOS Stream 8 - oVirt 4.5 - powertools Rocky Linux 8 - PowerTools So I kicked-off the oVirt deployment with: "hosted-engine --deploy --4 --ansible-extra-vars=he_offline_deployment=true". I used "--ansible-extra-vars=he_offline_deployment=true" because without that flag I was getting "DNF timout" issues (see my previous post `Local (Deployment) VM Can't Reach "centos-ceph-pacific" Repo`). I answer the defaults to all of questions the script asked, or entered the deployment-relevant answers where appropriate. In doing this I double-checked every answer before hitting <Enter>. Everything progressed smoothly until the deployment reached the "Wait for the host to be up" task... which then hung for more than 30 minutes before failing. From the ovirt-hosted-engine-setup... log file: - 2022-10-20 17:54:26,285+1100 ERROR otopi.ovirt_hosted_engine_setup.ansible_utils ansible_utils._process_output:113 fatal: [localhost]: FAILED! => {"changed": false, "msg": "Host is not up, please check logs, perhaps also on the engine machine"} I checked the following log files and found all of the relevant ERROR lines, then checked several 10s of proceeding and succeeding lines trying to determine what was going wrong, but I could not determine anything. - ovirt-hosted-engine-setup... - ovirt-hosted-engine-setup-ansible-bootstrap_local_vm... - ovirt-hosted-engine-setup-ansible-final_clean... - not really relevant, I believe I can include the log files (or the relevant parts of the log files) if people want - but that are very large: several 100 kilobytes each. I also googled "oVirt Host is not up" and found several entries, but after reading them all the most relevant seems to be a thread from these mailing list: `Install of RHV 4.4 failing - "Host is not up, please check logs, perhaps also on the engine machine"` - but this seems to be talking about an upgrade and I didn't gleam anything useful from it - I could, of course, be wrong about that. So my questions are: - Where else should I be looking (ie other log files, etc, and possible where to find them)? - Does anyone have any idea why this isn't working? - Does anyone have a work-around (including a completely manual process to get things working - I don't mind working in the CLI with virsh, etc)? - What am I doing wrong? Please, I'm really stumped with this, and I really do need help. Cheers Dulux-Oz _______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-leave@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: users@ovirt.org <https://lists.ovirt.org/archives/list/<a href=>/message/VBCK7C3Z22K7VZVXDBY5MVZPG5YGFHD6/">https://lists.ovirt.org/archives/list/users@ovirt.org/message/VBCK7C3Z22K7VZ...

Question: Are the hosted-engine flags "--ansible-extra-vars=he_offline_deployment" & "--ansible-extra-vars=he_pause_host" mutually exclusive? The reason I ask is that when I run "hosted-engine --deploy --4 --ansible-extra-vars=he_offline_deployment=true --ansible-extra-vars=he_pause_host=true" the script runs into the "DNF Timeout" issue I mentioned in the `Local (Deployment) VM Can't Reach "centos-ceph-pacific" Repo`, and running "hosted-engine --deploy --4 --ansible-extra-vars=he_offline_deployment=true" doesn't encounter that issue. In other words, I can't do what you suggest (thank you for the suggestion, btw) because I either run into the "DNF Timeout" issue (and thus don't get anywhere near being able to log into the the engine), or I can't set "--ansible-extra-vars=he_pause_host" to true. Any ideas? Cheers Dulux-Oz

OK, so, with all the tooing-and-frowing things stand as follows (@03:15UTC 25-Oct-2022): - I managed to solve the "DNF Timeout" issue (see my post "Local (Deployment) VM Can't Reach "centos-ceph-pacific" Repo") and so simplified the deployment command to `hosted-engine --deploy`. Unfortunately this still results in a "Host is not up" error, with the logs as per before. - As mentioned elsewhere in this thread I uploaded the (previous) logs to Dropbox along with a couple of other relevant(?) files: https://www.dropbox.com/sh/eymwdy8hzn3sa7z/AACscSP2eaFfoiN-QzyeEVfaa?dl=0 - I followed the suggestion of ajude.pereira (see post in this thread) but this did not resolve the issue. - As per one of my other posts in this thread, digging into the logs further revealed this issue: "Failed to authenticate session with host 'ovirt_node_1.mynet.local': SSH authentication to 'root(a)ovirt_node_1.mynet.local' failed. Please verify provided credentials. Make sure key is authorized at host" - I also did a `hosted-engine --deploy --ansible-extra-vars=he_pause_host=true` (as per the suggestion of Konstantin - see post in this thread) and tried to work out why ssh wasn't working. I ssh'd into the deployment VM and then attempted to ssh back into the deployment host (ie `ssh root@ovirt_node_1.mynet.local`). While I could connect, I was asked for the root's password. I was under the impression that this was supposed to be a "password-less" operation. As I do not provide the root@ovirt_node_1.mynet.local password anywhere in the deployment script, I suspect that this is why I'm getting the "Host is not up" error. - To reiterate: the host'd sshd_config file is configured as per the oVirt documentation. So am I wrong in my understanding of the password-less ssh-nature of the situation and how the deployment script is supposed to work? Also, does *anyone* have any pointers, suggestions, or can otherwise help me out - thanks. Cheers Dulux-Oz

On Tue, Oct 25, 2022 at 6:27 AM Matthew J Black <matthew@peregrineit.net> wrote:
OK, so, with all the tooing-and-frowing things stand as follows (@03:15UTC 25-Oct-2022):
- I managed to solve the "DNF Timeout" issue (see my post "Local (Deployment) VM Can't Reach "centos-ceph-pacific" Repo") and so simplified the deployment command to `hosted-engine --deploy`. Unfortunately this still results in a "Host is not up" error, with the logs as per before.
- As mentioned elsewhere in this thread I uploaded the (previous) logs to Dropbox along with a couple of other relevant(?) files: https://www.dropbox.com/sh/eymwdy8hzn3sa7z/AACscSP2eaFfoiN-QzyeEVfaa?dl=0
- I followed the suggestion of ajude.pereira (see post in this thread) but this did not resolve the issue.
- As per one of my other posts in this thread, digging into the logs further revealed this issue: "Failed to authenticate session with host 'ovirt_node_1.mynet.local': SSH authentication to 'root(a)ovirt_node_1.mynet.local' failed. Please verify provided credentials. Make sure key is authorized at host"
- I also did a `hosted-engine --deploy --ansible-extra-vars=he_pause_host=true` (as per the suggestion of Konstantin - see post in this thread) and tried to work out why ssh wasn't working. I ssh'd into the deployment VM and then attempted to ssh back into the deployment host (ie `ssh root@ovirt_node_1.mynet.local`). While I could connect, I was asked for the root's password.
Good.
I was under the impression that this was supposed to be a "password-less" operation.
It should. At this point, the operation that is attempted and which is failing with the error you see in engine.log ("Failed to authenticate session"), is done using Java code, using the Java library apache-sshd, not the command line ssh. Some of the relevant code is here: https://github.com/oVirt/ovirt-engine/blob/master/backend/manager/modules/bl... I do not know this code well, sorry, nor the specifics of apache-sshd vs openssh (and there are such "specifics", as can easily be seen by looking at the engine git log).
As I do not provide the root@ovirt_node_1.mynet.local password anywhere in the deployment script, I suspect that this is why I'm getting the "Host is not up" error.
- To reiterate: the host'd sshd_config file is configured as per the oVirt documentation.
So am I wrong in my understanding of the password-less ssh-nature of the situation and how the deployment script is supposed to work?
I think this should work more or less like this: After running engine-setup, and when the engine is already up, we fetch the public key of the engine from it, and store it in your authorized_keys file. This is done here: https://github.com/oVirt/ovirt-ansible-collection/blob/master/roles/hosted_e... - name: Set Engine public key as authorized key without validating the TLS/SSL certificates I do see this in your log in dropbox. Do you see /root/.ssh/authorized_keys on the host (with a timestamp similar to the log line)? If so, you can try this, from the engine VM: ssh -v -i /etc/pki/ovirt-engine/keys/engine_id_rsa ovirt_node_1.mynet.local If this does not work, you can continue debugging this until you manage to understand/fix. Perhaps check sshd config etc. If it does work, it means the issue might be due to incompatibility between apache-sshd and openssh and/or the configuration.
Also, does *anyone* have any pointers, suggestions, or can otherwise help me out - thanks.
At this point, you should be able to log into the admin UI (the pause message provides a link) and try to manually add the host. It seems like this didn't work for you. This is because "host_result_up_check" is "failed", and we pause only if it succeeded and the host is returned with status "non_operational". Feel free to create an issue to make the code pause also if "host_result_up_check" is "failed" - not sure why we do not, perhaps we did have a reason. Anyway, you can force the code to pause after trying to add the host but before checking if this worked, by passing "--ansible-extra-vars=he_pause_host=true". You can also check/share more of engine.log - there might be more information prior to the failure (but as I said, I do not know this code well). You can try running sshd (the server) with debug info and check its own log - the issue might be due to incompatible keys on one or both of the sides, or something like that. Sorry that I do not remember if you wrote this before - is this your first attempt to install oVirt? If so, perhaps try first to start with a clean host, without any custom configuration (e.g. of sshd), and see if this works for you. If you do have access to a successful setup, you can more easily compare. Good luck and best regards, -- Didi

On 25/10/2022 17:23, Yedidyah Bar David wrote:
If so, you can try this, from the engine VM: ssh -v -i /etc/pki/ovirt-engine/keys/engine_id_rsa ovirt_node_1.mynet.local
Tried that
If it does work, it means the issue might be due to incompatibility between apache-sshd and openssh and/or the configuration.
And it did - so now what (logs will be uploaded in a few hours (after I get some sleep :-) )? Also, with the Deployment VM being on 192.168.222.77, and the (headless) host being on 172.16.1.100 (and my Windows PC (the one with an actual GUI) being on 172.16.1.50, how do I log into the Cockpit on the Deployment VM - ie there is no route, apart from a static routed added to the PC pointing towards the host (route add 192.168.222.0/24 172.16.1.100) - and that still doesn't allow me to https into the deployment VM - what am I missing? And finally, what's happening from oVirt's side re: the recent cockpit issue with modern browsers (ie Firefox and Chrome) being refused connection - is/does this need to be updated at all? Cheers Dulux-Oz -- This email has been checked for viruses by Avast antivirus software. www.avast.com

Sorry it took me a few days, but the (new) logs have been placed on DropBox (the existing URL, above, should still work) Cheers Dulux-Oz

Hi All, Long story short, I just tried to do a `hosted-engine --deploy` on a brand, new "out-of-the-box box", following the oVirt doco *exactly*, and while I got past my "Host is not up" issue, but almost exactly afterwards got this in my console (I've included the couple of lines leading up to the "Host is no up" step): ~~~ [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Obtain SSO token using username/password credentials] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Wait for the host to be up] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Notify the user about a failure] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Set host_id] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Collect error events from the Engine] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Generate the error message from the engine events] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Notify with error description] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Notify with generic error] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Let the user connect to the bootstrap engine to manually fix host configuration] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : include_tasks] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Create temporary lock file] [ INFO ] changed: [localhost -> localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Pause execution until /tmp/ansible.volt5pvv_he_setup_lock is removed, delete it once ready to proceed] ~~~ I didn't ask for script to pause, so I need to ask: Is this normal? I'm about to have a look at the logs, but its late here at the moment so I wanted to get this up on the mailing lists so I don't loose too much time while I'm asleep and everyone else is awake (& vice-versa). Cheers Dulux-Oz

On 1. 11. 2022, at 11:46, Matthew J Black <matthew@peregrineit.net> wrote:
Hi All,
Long story short, I just tried to do a `hosted-engine --deploy` on a brand, new "out-of-the-box box", following the oVirt doco *exactly*, and while I got past my "Host is not up" issue, but almost exactly afterwards got this in my console (I've included the couple of lines leading up to the "Host is no up" step):
~~~ [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Obtain SSO token using username/password credentials] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Wait for the host to be up] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Notify the user about a failure] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Set host_id] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Collect error events from the Engine] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Generate the error message from the engine events] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Notify with error description] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Notify with generic error] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Let the user connect to the bootstrap engine to manually fix host configuration] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : include_tasks] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Create temporary lock file] [ INFO ] changed: [localhost -> localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Pause execution until /tmp/ansible.volt5pvv_he_setup_lock is removed, delete it once ready to proceed] ~~~
I didn't ask for script to pause, so I need to ask: Is this normal?
No. I guess you did ask for that, maybe by mistake. The default is not to pause. can you get to webadmin now and confirm th host state? and/or just remove that file and let it continue... Thanks, michal
I'm about to have a look at the logs, but its late here at the moment so I wanted to get this up on the mailing lists so I don't loose too much time while I'm asleep and everyone else is awake (& vice-versa).
Cheers
Dulux-Oz _______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-leave@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/AZ7W7J6ANA3MEL...

On Tue, Nov 1, 2022 at 3:14 PM Michal Skrivanek <mskrivan@redhat.com> wrote:
On 1. 11. 2022, at 11:46, Matthew J Black <matthew@peregrineit.net> wrote:
Hi All,
Long story short, I just tried to do a `hosted-engine --deploy` on a brand, new "out-of-the-box box", following the oVirt doco *exactly*, and while I got past my "Host is not up" issue, but almost exactly afterwards got this in my console (I've included the couple of lines leading up to the "Host is no up" step):
~~~ [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Obtain SSO token using username/password credentials] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Wait for the host to be up] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Notify the user about a failure] [ INFO ] skipping: [localhost]
Skipped, so this means that host_result_up_check (the check itself) didn't fail.
[ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Set host_id] [ INFO ] ok: [localhost]
This means that we entered the block "Emit error messages about the failure", which means all of the 3 "and-ed" conditions in its "when:" were true. Sadly, this (the block) isn't logged in itself - see [1] (and [2], a simple attempt at a workaround that I eventually gave up on). This means that we did get a host in the result, and that its status is "non_operational". [1] https://github.com/ansible/ansible/issues/33581 [2] https://gerrit.ovirt.org/c/ovirt-hosted-engine-setup/+/115775
[ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Collect error events from the Engine] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Generate the error message from the engine events] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Notify with error description] [ INFO ] ok: [localhost]
This means that the above tasks did manage to generate an "error_description", but that for some reason it was not output to your console. This is quite weird, I have no idea how this happened. I am confident I tested it back then [3], and also can find examples (very few) where it worked for others [4]. Incidentally, [4] does not include the task name "Notify with error description", and neither google finds it anywhere, but your output above does include it. [4] does include the actual text, but your output does not. Perhaps this is a change in ansible since then or something similar. Would you like to create a github issue in [5] about it? Something like "HE deployment gets stuck without details" or whatever, and add relevant parts of the logs. Thanks. [3] https://github.com/oVirt/ovirt-ansible-collection/pull/181 [4] https://lists.ovirt.org/archives/list/users@ovirt.org/thread/UZHD5KCPMKXOSPF... [5] https://github.com/oVirt/ovirt-ansible-collection
[ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Notify with generic error] [ INFO ] skipping: [localhost]
(This is another indication that we did manage to generate an error message)
[ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Let the user connect to the bootstrap engine to manually fix host configuration] [ INFO ] ok: [localhost]
Again - the actual text is missing.
[ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : include_tasks] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Create temporary lock file] [ INFO ] changed: [localhost -> localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Pause execution until /tmp/ansible.volt5pvv_he_setup_lock is removed, delete it once ready to proceed] ~~~
I didn't ask for script to pause, so I need to ask: Is this normal?
No. I guess you did ask for that, maybe by mistake. The default is not to pause.
This isn't the case here - it paused due to a problem, but didn't emit details. Can you please check the logs in /var/log/ovirt-hosted-engine-setup? You most likely should be able to find the messages there, including the URL to the web admin if needed.
can you get to webadmin now and confirm th host state? and/or just remove that file and let it continue...
If you do that, it will shortly thereafter fail in the following task "Check if the host is up".
Thanks, michal
I'm about to have a look at the logs, but its late here at the moment so I wanted to get this up on the mailing lists so I don't loose too much time while I'm asleep and everyone else is awake (& vice-versa).
Good luck, sorry for all the obstacles you bump into, and best regards, -- Didi

OK, so as I said I was going to do I've now gone through the logs. I've place the log files into DropBox (https://www.dropbox.com/sh/eymwdy8hzn3sa7z/AACscSP2eaFfoiN-QzyeEVfaa?dl=0) There was only one significant part of the logs (at least that what it appears to me) and I've included that extract below: ovirt-hosted-engine-setup-ansible-bootstrap_local_vm-...log Extract ~~~ 2022-11-01 21:34:57,395+1100 INFO ansible task start {'status': 'OK', 'ansible_type': 'task', 'ansible_playbook': '/usr/share/ovirt-hosted-engine-setup/he_ansible/trigger_role.yml', 'ansible_task': 'ovirt.ovirt.hosted_engine_setup : Notify the user about a failure'} 2022-11-01 21:34:57,395+1100 DEBUG ansible on_any args TASK: ovirt.ovirt.hosted_engine_setup : Notify the user about a failure kwargs is_conditional:False 2022-11-01 21:34:57,396+1100 DEBUG ansible on_any args localhost TASK: ovirt.ovirt.hosted_engine_setup : Notify the user about a failure kwargs 2022-11-01 21:34:57,875+1100 INFO ansible skipped {'status': 'SKIPPED', 'ansible_type': 'task', 'ansible_playbook': '/usr/share/ovirt-hosted-engine-setup/he_ansible/trigger_role.yml', 'ansible_task': 'Notify the user about a failure', 'ansible_host': 'localhost'} 2022-11-01 21:34:57,876+1100 DEBUG ansible on_any args <ansible.executor.task_result.TaskResult object at 0x7f0b21370ee0> kwargs 2022-11-01 21:34:58,359+1100 INFO ansible task start {'status': 'OK', 'ansible_type': 'task', 'ansible_playbook': '/usr/share/ovirt-hosted-engine-setup/he_ansible/trigger_role.yml', 'ansible_task': 'ovirt.ovirt.hosted_engine_setup : Set host_id'} 2022-11-01 21:34:58,359+1100 DEBUG ansible on_any args TASK: ovirt.ovirt.hosted_engine_setup : Set host_id kwargs is_conditional:False 2022-11-01 21:34:58,360+1100 DEBUG ansible on_any args localhost TASK: ovirt.ovirt.hosted_engine_setup : Set host_id kwargs 2022-11-01 21:34:58,844+1100 DEBUG var changed: host "localhost" var "host_id" type "<class 'ansible.utils.unsafe_proxy.AnsibleUnsafeText'>" value: ""eb33e62a-2929-499f-80de-b7ac38a075f5"" 2022-11-01 21:34:58,844+1100 INFO ansible ok {'status': 'OK', 'ansible_type': 'task', 'ansible_playbook': '/usr/share/ovirt-hosted-engine-setup/he_ansible/trigger_role.yml', 'ansible_host': 'localhost', 'ansible_task': 'Set host_id', 'task_duration': 0} 2022-11-01 21:34:58,844+1100 DEBUG ansible on_any args <ansible.executor.task_result.TaskResult object at 0x7f0b214c4a00> kwargs 2022-11-01 21:34:59,288+1100 INFO ansible task start {'status': 'OK', 'ansible_type': 'task', 'ansible_playbook': '/usr/share/ovirt-hosted-engine-setup/he_ansible/trigger_role.yml', 'ansible_task': 'ovirt.ovirt.hosted_engine_setup : Collect error events from the Engine'} 2022-11-01 21:34:59,289+1100 DEBUG ansible on_any args TASK: ovirt.ovirt.hosted_engine_setup : Collect error events from the Engine kwargs is_conditional:False 2022-11-01 21:34:59,290+1100 DEBUG ansible on_any args localhost TASK: ovirt.ovirt.hosted_engine_setup : Collect error events from the Engine kwargs 2022-11-01 21:35:00,157+1100 INFO ansible ok {'status': 'OK', 'ansible_type': 'task', 'ansible_playbook': '/usr/share/ovirt-hosted-engine-setup/he_ansible/trigger_role.yml', 'ansible_host': 'localhost', 'ansible_task': 'Collect error events from the Engine', 'task_duration': 1} 2022-11-01 21:35:00,157+1100 DEBUG ansible on_any args <ansible.executor.task_result.TaskResult object at 0x7f0b214ae760> kwargs 2022-11-01 21:35:00,625+1100 DEBUG var changed: host "localhost" var "error_events" type "<class 'dict'>" value: "{ "changed": false, "failed": false, "ovirt_events": [ { "cluster": { "href": "/ovirt-engine/api/clusters/c44e2594-989d-4f1e-8308-feec46918d67", "id": "c44e2594-989d-4f1e-8308-feec46918d67", "name": "my_cluster_1" }, "code": 532, "custom_id": -1, "description": "Used memory of host ovirt_node_1.mynet.local in cluster my_cluster_1 [100%] exceeded defined threshold [95%].", "flood_rate": 0, "host": { "href": "/ovirt-engine/api/hosts/eb33e62a-2929-499f-80de-b7ac38a075f5", "id": "eb33e62a-2929-499f-80de-b7ac38a075f5", "name": "ovirt_node_1.mynet.local" }, "href": "/ovirt-engine/api/events/142", "id": "142", "index": 142, "origin": "oVirt", "severity": "warning", "time": "2022-11-01 21:34:57.640000+11:00" }, { "cluster": { "href": "/ovirt-engine/api/clusters/c44e2594-989d-4f1e-8308-feec46918d67", "id": "c44e2594-989d-4f1e-8308-feec46918d67", "name": "my_cluster_1" }, "code": 519, "correlation_id": "65a04e79", "custom_id": -1, "description": "Host ovirt_node_1.mynet.local does not comply with the cluster my_cluster_1 networks, the following networks are missing on host: 'ovirtmgmt'", "flood_rate": 0, "host": { "href": "/ovirt-engine/api/hosts/eb33e62a-2929-499f-80de-b7ac38a075f5", "id": "eb33e62a-2929-499f-80de-b7ac38a075f5", "name": "ovirt_node_1.mynet.local" }, "href": "/ovirt-engine/api/events/140", "id": "140", "index": 140, "origin": "oVirt", "severity": "warning", "time": "2022-11-01 21:34:57.404000+11:00", "user": { "name": "SYSTEM" } }, { "cluster": { "href": "/ovirt-engine/api/clusters/c44e2594-989d-4f1e-8308-feec46918d67", "id": "c44e2594-989d-4f1e-8308-feec46918d67", "name": "my_cluster_1" }, "code": 505, "correlation_id": "8a77e33b-64f8-409b-abe1-78a37ae6df4c", "custom_id": -1, "description": "Host ovirt_node_1.mynet.local installation failed. Failed to configure management network on the host.", "flood_rate": 0, "host": { "href": "/ovirt-engine/api/hosts/eb33e62a-2929-499f-80de-b7ac38a075f5", "id": "eb33e62a-2929-499f-80de-b7ac38a075f5", "name": "ovirt_node_1.mynet.local" }, "href": "/ovirt-engine/api/events/137", "id": "137", "index": 137, "origin": "oVirt", "severity": "error", "time": "2022-11-01 21:34:51.506000+11:00", "user": { "href": "/ovirt-engine/api/users/66aaaec1-29f1-4e4f-a8c6-2a9c7f3319c2", "id": "66aaaec1-29f1-4e4f-a8c6-2a9c7f3319c2", "name": "admin@ovirt@internalkeycloak-authz" } }, { "code": 1120, "custom_id": -1, "description": "Failed to configure management network on host ovirt_node_1.mynet.local due to setup networks failure.", "flood_rate": 0, "host": { "href": "/ovirt-engine/api/hosts/eb33e62a-2929-499f-80de-b7ac38a075f5", "id": "eb33e62a-2929-499f-80de-b7ac38a075f5", "name": "ovirt_node_1.mynet.local" }, "href": "/ovirt-engine/api/events/136", "id": "136", "index": 136, "origin": "oVirt", "severity": "error", "time": "2022-11-01 21:34:51.413000+11:00" }, { "code": 10802, "custom_id": -1, "description": "VDSM ovirt_node_1.mynet.local command HostSetupNetworksVDS failed: Internal JSON-RPC error: {'reason': \"\\ndesired\\n=======\\n---\\nname: bond_1\\ntype: bond\\nstate: up\\nipv4:\\n enabled: false\\nipv6:\\n enabled: false\\nlink-aggregation:\\n mode: 802.3ad\\n options:\\n downdelay: 1000\\n miimon: 1000\\n updelay: 1000\\n port:\\n - eno1\\nmac-address: 3C:EC:EF:83:77:4C\\nmtu: 1500\\n\\ncurrent\\n=======\\n---\\nname: bond_1\\ntype: bond\\nstate: up\\naccept-all-mac-addresses: false\\nethtool:\\n feature:\\n esp-hw-offload: false\\n esp-tx-csum-hw-offload: false\\n highdma: true\\n rx-gro: true\\n rx-gro-list: false\\n rx-lro: false\\n rx-udp-gro-forwarding: false\\n rx-vlan-filter: true\\n rx-vlan-hw-parse: true\\n tx-checksum-ip-generic: true\\n tx-esp-segmentation: false\\n tx-generic-segmentation: true\\n tx-gre-csum-segmentation: true\\n tx-gre-segmentation: true\\n tx-gso-list: false\\n tx-ipxip4-segmentation: true\\n tx-ipxip6-segmentation: true\\n tx-nocache-copy: false\\n tx-scatter-gather-fraglist: false\\n tx-sctp-segmentation: false\\n tx-tcp-ecn-segmentation: true\\n tx-tcp-mangleid-segmentation: true\\n tx-tcp-segmentation: true\\n tx-tcp6-segmentation: true\\n tx-udp-segmentation: true\\n tx-udp_tnl-csum-segmentation: true\\n tx-udp_tnl-segmentation: true\\nipv4:\\n enabled: false\\nipv6:\\n enabled: false\\nlink-aggregation:\\n mode: 802.3ad\\n options:\\n ad_actor_sys_prio: 65535\\n ad_actor_system: 00:00:00:00:00:00\\n ad_select: stable\\n ad_user_port_key: 0\\n all_slaves_active: dropped\\n arp_all_targets: any\\n arp_interval: 0\\n arp_ip_target: ''\\n arp_validate: none\\n downdelay: 1000\\n lacp_rate: slow\\n miimon: 1000\\n min_links: 0\\n updelay: 1000\\n use_carrier: true\\n xmit_hash_policy: layer2\\n port:\\n - eno1\\n - eno2\\nlldp:\\n enabled: false\\ nmac-address: 3C:EC:EF:83:77:4C\\nmtu: 1500\\n\\ndifference\\n==========\\n--- desired\\n+++ current\\n@@ -2,6 +2,36 @@\\n name: bond_1\\n type: bond\\n state: up\\n+accept-all-mac-addresses: false\\n+ethtool:\\n+ feature:\\n+ esp-hw-offload: false\\n+ esp-tx-csum-hw-offload: false\\n+ highdma: true\\n+ rx-gro: true\\n+ rx-gro-list: false\\n+ rx-lro: false\\n+ rx-udp-gro-forwarding: false\\n+ rx-vlan-filter: true\\n+ rx-vlan-hw-parse: true\\n+ tx-checksum-ip-generic: true\\n+ tx-esp-segmentation: false\\n+ tx-generic-segmentation: true\\n+ tx-gre-csum-segmentation: true\\n+ tx-gre-segmentation: true\\n+ tx-gso-list: false\\n+ tx-ipxip4-segmentation: true\\n+ tx-ipxip6-segmentation: true\\n+ tx-nocache-copy: false\\n+ tx-scatter-gather-fraglist: false\\n+ tx-sctp-segmentation: false\\n+ tx-tcp-ecn-segmentation: true\\n+ tx-tcp-mangleid-segmentation: true\\n+ tx-tcp-segmentation: true\\n+ tx-tcp6-segmentation: tr ue\\n+ tx-udp-segmentation: true\\n+ tx-udp_tnl-csum-segmentation: true\\n+ tx-udp_tnl-segmentation: true\\n ipv4:\\n enabled: false\\n ipv6:\\n@@ -9,10 +39,26 @@\\n link-aggregation:\\n mode: 802.3ad\\n options:\\n+ ad_actor_sys_prio: 65535\\n+ ad_actor_system: 00:00:00:00:00:00\\n+ ad_select: stable\\n+ ad_user_port_key: 0\\n+ all_slaves_active: dropped\\n+ arp_all_targets: any\\n+ arp_interval: 0\\n+ arp_ip_target: ''\\n+ arp_validate: none\\n downdelay: 1000\\n+ lacp_rate: slow\\n miimon: 1000\\n+ min_links: 0\\n updelay: 1000\\n+ use_carrier: true\\n+ xmit_hash_policy: layer2\\n port:\\n - eno1\\n+ - eno2\\n+lldp:\\n+ enabled: false\\n mac-address: 3C:EC:EF:83:77:4C\\n mtu: 1500\\n\\n\"}", "flood_rate": 0, "host": { "href": "/ovirt-engine/api/hosts/eb33e62a-2929-499f-80de-b7ac38a075f5", "id": "eb33e62a-2929-499f-80de-b7ac38a075f5", "name": "ovirt_node_1.mynet.local" }, "href": "/ovirt-engine/api/events/135", "id": "135", "index": 135, "origin": "oVirt", "severity": "error", "time": "2022-11-01 21:34:51.387000+11:00" }, { "cluster": { "href": "/ovirt-engine/api/clusters/c44e2594-989d-4f1e-8308-feec46918d67", "id": "c44e2594-989d-4f1e-8308-feec46918d67", "name": "my_cluster_1" }, "code": 553, "correlation_id": "8a77e33b-64f8-409b-abe1-78a37ae6df4c", "custom_id": -1, "description": "Installing Host ovirt_node_1.mynet.local. Check for LVM filter configuration error: Cannot configure LVM filter on host, please run: vdsm-tool config-lvm-filter.", "flood_rate": 0, "host": { "href": "/ovirt-engine/api/hosts/eb33e62a-2929-499f-80de-b7ac38a075f5", "id": "eb33e62a-2929-499f-80de-b7ac38a075f5", "name": "ovirt_node_1.mynet.local" }, "href": "/ovirt-engine/api/events/59", "id": "59", "index": 59, "origin": "oVirt", "severity": "error", "time": "2022-11-01 21:32:19.692000+11:00" }, { "code": 9000, "custom_id": -1, "description": "Failed to verify Power Management configuration for Host ovirt_node_1.mynet.local.", "flood_rate": 0, "host": { "href": "/ovirt-engine/api/hosts/eb33e62a-2929-499f-80de-b7ac38a075f5", "id": "eb33e62a-2929-499f-80de-b7ac38a075f5", "name": "ovirt_node_1.mynet.local" }, "href": "/ovirt-engine/api/events/13", "id": "13", "index": 13, "origin": "oVirt", "severity": "alert", "time": "2022-11-01 21:31:44.645000+11:00" }, { "code": 11291, "custom_id": -1, "description": "Update to network ovirtmgmt was not applied to virtual network interfaces [<UNKNOWN>]. The actual configuration on the interfaces may differ from the displayed one.", "flood_rate": 0, "href": "/ovirt-engine/api/events/10", "id": "10", "index": 10, "origin": "oVirt", "severity": "alert", "time": "2022-11-01 21:31:35.606000+11:00" } ] }" 2022-11-01 21:35:00,625+1100 INFO ansible task start {'status': 'OK', 'ansible_type': 'task', 'ansible_playbook': '/usr/share/ovirt-hosted-engine-setup/he_ansible/trigger_role.yml', 'ansible_task': 'ovirt.ovirt.hosted_engine_setup : Generate the error message from the engine events'} 2022-11-01 21:35:00,625+1100 DEBUG ansible on_any args TASK: ovirt.ovirt.hosted_engine_setup : Generate the error message from the engine events kwargs is_conditional:False 2022-11-01 21:35:00,626+1100 DEBUG ansible on_any args localhost TASK: ovirt.ovirt.hosted_engine_setup : Generate the error message from the engine events kwargs 2022-11-01 21:35:01,125+1100 DEBUG var changed: host "localhost" var "error_description" type "<class 'ansible.utils.unsafe_proxy.AnsibleUnsafeText'>" value: "" code 505: Host ovirt_node_1.mynet.local installation failed. Failed to configure management network on the host., code 519: Host ovirt_node_1.mynet.local does not comply with the cluster my_cluster_1 networks, the following networks are missing on host: 'ovirtmgmt', code 532: Used memory of host ovirt_node_1.mynet.local in cluster my_cluster_1 [100%] exceeded defined threshold [95%]., code 553: Installing Host ovirt_node_1.mynet.local. Check for LVM filter configuration error: Cannot configure LVM filter on host, please run: vdsm-tool config-lvm-filter., code 1120: Failed to configure management network on host ovirt_node_1.mynet.local due to setup networks failure., code 9000: Failed to verify Power Management configuration for Host ovirt_node_1.mynet.local., code 10802: VDSM ovirt_node_1.mynet.local comma nd HostSetupNetworksVDS failed: Internal JSON-RPC error: {'reason': \"\\ndesired\\n=======\\n---\\nname: bond_1\\ntype: bond\\nstate: up\\nipv4:\\n enabled: false\\nipv6:\\n enabled: false\\nlink-aggregation:\\n mode: 802.3ad\\n options:\\n downdelay: 1000\\n miimon: 1000\\n updelay: 1000\\n port:\\n - eno1\\nmac-address: 3C:EC:EF:83:77:4C\\nmtu: 1500\\n\\ncurrent\\n=======\\n---\\nname: bond_1\\ntype: bond\\nstate: up\\naccept-all-mac-addresses: false\\nethtool:\\n feature:\\n esp-hw-offload: false\\n esp-tx-csum-hw-offload: false\\n highdma: true\\n rx-gro: true\\n rx-gro-list: false\\n rx-lro: false\\n rx-udp-gro-forwarding: false\\n rx-vlan-filter: true\\n rx-vlan-hw-parse: true\\n tx-checksum-ip-generic: true\\n tx-esp-segmentation: false\\n tx-generic-segmentation: true\\n tx-gre-csum-segmentation: true\\n tx-gre-segmentation: true\\n tx-gso-list: false\\n tx-ipxip4-segmentation: true\\n tx-ipxip6-segmentation: true\\n tx-nocache-copy: false\\n tx-scatter-gather-fraglist: false\\n tx-sctp-segmentation: false\\n tx-tcp-ecn-segmentation: true\\n tx-tcp-mangleid-segmentation: true\\n tx-tcp-segmentation: true\\n tx-tcp6-segmentation: true\\n tx-udp-segmentation: true\\n tx-udp_tnl-csum-segmentation: true\\n tx-udp_tnl-segmentation: true\\nipv4:\\n enabled: false\\nipv6:\\n enabled: false\\nlink-aggregation:\\n mode: 802.3ad\\n options:\\n ad_actor_sys_prio: 65535\\n ad_actor_system: 00:00:00:00:00:00\\n ad_select: stable\\n ad_user_port_key: 0\\n all_slaves_active: dropped\\n arp_all_targets: any\\n arp_interval: 0\\n arp_ip_target: ''\\n arp_validate: none\\n downdelay: 1000\\n lacp_rate: slow\\n miimon: 1000\\n min_links: 0\\n updelay: 1000\\n use_carrier: true\\n xmit_hash_policy: layer2\\n port:\\n - eno1\\n - eno2\\nlldp:\\n enabled: false\\nmac-address: 3C:EC:EF:83:77:4C\\nmtu: 1500\\n\\ndifference\\n ==========\\n--- desired\\n+++ current\\n@@ -2,6 +2,36 @@\\n name: bond_1\\n type: bond\\n state: up\\n+accept-all-mac-addresses: false\\n+ethtool:\\n+ feature:\\n+ esp-hw-offload: false\\n+ esp-tx-csum-hw-offload: false\\n+ highdma: true\\n+ rx-gro: true\\n+ rx-gro-list: false\\n+ rx-lro: false\\n+ rx-udp-gro-forwarding: false\\n+ rx-vlan-filter: true\\n+ rx-vlan-hw-parse: true\\n+ tx-checksum-ip-generic: true\\n+ tx-esp-segmentation: false\\n+ tx-generic-segmentation: true\\n+ tx-gre-csum-segmentation: true\\n+ tx-gre-segmentation: true\\n+ tx-gso-list: false\\n+ tx-ipxip4-segmentation: true\\n+ tx-ipxip6-segmentation: true\\n+ tx-nocache-copy: false\\n+ tx-scatter-gather-fraglist: false\\n+ tx-sctp-segmentation: false\\n+ tx-tcp-ecn-segmentation: true\\n+ tx-tcp-mangleid-segmentation: true\\n+ tx-tcp-segmentation: true\\n+ tx-tcp6-segmentation: true\\n+ tx-udp-segmentation: true\\n+ tx-udp_tnl-csum-seg mentation: true\\n+ tx-udp_tnl-segmentation: true\\n ipv4:\\n enabled: false\\n ipv6:\\n@@ -9,10 +39,26 @@\\n link-aggregation:\\n mode: 802.3ad\\n options:\\n+ ad_actor_sys_prio: 65535\\n+ ad_actor_system: 00:00:00:00:00:00\\n+ ad_select: stable\\n+ ad_user_port_key: 0\\n+ all_slaves_active: dropped\\n+ arp_all_targets: any\\n+ arp_interval: 0\\n+ arp_ip_target: ''\\n+ arp_validate: none\\n downdelay: 1000\\n+ lacp_rate: slow\\n miimon: 1000\\n+ min_links: 0\\n updelay: 1000\\n+ use_carrier: true\\n+ xmit_hash_policy: layer2\\n port:\\n - eno1\\n+ - eno2\\n+lldp:\\n+ enabled: false\\n mac-address: 3C:EC:EF:83:77:4C\\n mtu: 1500\\n\\n\"}, "" 2022-11-01 21:35:01,125+1100 INFO ansible ok {'status': 'OK', 'ansible_type': 'task', 'ansible_playbook': '/usr/share/ovirt-hosted-engine-setup/he_ansible/trigger_role.yml', 'ansible_host': 'localhost', 'ansible_task': 'Generate the error message from the engine events', 'task_duration': 0} 2022-11-01 21:35:01,125+1100 DEBUG ansible on_any args <ansible.executor.task_result.TaskResult object at 0x7f0b21370ee0> kwargs 2022-11-01 21:35:01,583+1100 INFO ansible task start {'status': 'OK', 'ansible_type': 'task', 'ansible_playbook': '/usr/share/ovirt-hosted-engine-setup/he_ansible/trigger_role.yml', 'ansible_task': 'ovirt.ovirt.hosted_engine_setup : Notify with error description'} 2022-11-01 21:35:01,583+1100 DEBUG ansible on_any args TASK: ovirt.ovirt.hosted_engine_setup : Notify with error description kwargs is_conditional:False 2022-11-01 21:35:01,584+1100 DEBUG ansible on_any args localhost TASK: ovirt.ovirt.hosted_engine_setup : Notify with error description kwargs 2022-11-01 21:35:02,032+1100 INFO ansible ok {'status': 'OK', 'ansible_type': 'task', 'ansible_playbook': '/usr/share/ovirt-hosted-engine-setup/he_ansible/trigger_role.yml', 'ansible_host': 'localhost', 'ansible_task': 'Notify with error description', 'task_duration': 0} 2022-11-01 21:35:02,032+1100 DEBUG ansible on_any args <ansible.executor.task_result.TaskResult object at 0x7f0b214ae790> kwargs 2022-11-01 21:35:02,511+1100 INFO ansible task start {'status': 'OK', 'ansible_type': 'task', 'ansible_playbook': '/usr/share/ovirt-hosted-engine-setup/he_ansible/trigger_role.yml', 'ansible_task': 'ovirt.ovirt.hosted_engine_setup : Notify with generic error'} 2022-11-01 21:35:02,511+1100 DEBUG ansible on_any args TASK: ovirt.ovirt.hosted_engine_setup : Notify with generic error kwargs is_conditional:False 2022-11-01 21:35:02,512+1100 DEBUG ansible on_any args localhost TASK: ovirt.ovirt.hosted_engine_setup : Notify with generic error kwargs 2022-11-01 21:35:02,954+1100 INFO ansible skipped {'status': 'SKIPPED', 'ansible_type': 'task', 'ansible_playbook': '/usr/share/ovirt-hosted-engine-setup/he_ansible/trigger_role.yml', 'ansible_task': 'Notify with generic error', 'ansible_host': 'localhost'} 2022-11-01 21:35:02,954+1100 DEBUG ansible on_any args <ansible.executor.task_result.TaskResult object at 0x7f0b21077520> kwargs 2022-11-01 21:35:03,401+1100 INFO ansible task start {'status': 'OK', 'ansible_type': 'task', 'ansible_playbook': '/usr/share/ovirt-hosted-engine-setup/he_ansible/trigger_role.yml', 'ansible_task': 'ovirt.ovirt.hosted_engine_setup : Let the user connect to the bootstrap engine to manually fix host configuration'} 2022-11-01 21:35:03,402+1100 DEBUG ansible on_any args TASK: ovirt.ovirt.hosted_engine_setup : Let the user connect to the bootstrap engine to manually fix host configuration kwargs is_conditional:False 2022-11-01 21:35:03,402+1100 DEBUG ansible on_any args localhost TASK: ovirt.ovirt.hosted_engine_setup : Let the user connect to the bootstrap engine to manually fix host configuration kwargs 2022-11-01 21:35:03,843+1100 INFO ansible ok {'status': 'OK', 'ansible_type': 'task', 'ansible_playbook': '/usr/share/ovirt-hosted-engine-setup/he_ansible/trigger_role.yml', 'ansible_host': 'localhost', 'ansible_task': 'Let the user connect to the bootstrap engine to manually fix host configuration', 'task_duration': 0} 2022-11-01 21:35:03,844+1100 DEBUG ansible on_any args <ansible.executor.task_result.TaskResult object at 0x7f0b214ae550> kwargs 2022-11-01 21:35:04,322+1100 INFO ansible task start {'status': 'OK', 'ansible_type': 'task', 'ansible_playbook': '/usr/share/ovirt-hosted-engine-setup/he_ansible/trigger_role.yml', 'ansible_task': 'ovirt.ovirt.hosted_engine_setup : include_tasks'} 2022-11-01 21:35:04,323+1100 DEBUG ansible on_any args TASK: ovirt.ovirt.hosted_engine_setup : include_tasks kwargs is_conditional:False 2022-11-01 21:35:04,323+1100 DEBUG ansible on_any args localhost TASK: ovirt.ovirt.hosted_engine_setup : include_tasks kwargs 2022-11-01 21:35:04,772+1100 INFO ansible ok {'status': 'OK', 'ansible_type': 'task', 'ansible_playbook': '/usr/share/ovirt-hosted-engine-setup/he_ansible/trigger_role.yml', 'ansible_host': 'localhost', 'ansible_task': '', 'task_duration': 0} 2022-11-01 21:35:04,772+1100 DEBUG ansible on_any args <ansible.executor.task_result.TaskResult object at 0x7f0b2154e700> kwargs 2022-11-01 21:35:04,798+1100 DEBUG ansible on_any args /usr/share/ansible/collections/ansible_collections/ovirt/ovirt/roles/hosted_engine_setup/tasks/pause_execution.yml (args={} vars={}): [localhost] kwargs 2022-11-01 21:35:05,260+1100 INFO ansible task start {'status': 'OK', 'ansible_type': 'task', 'ansible_playbook': '/usr/share/ovirt-hosted-engine-setup/he_ansible/trigger_role.yml', 'ansible_task': 'ovirt.ovirt.hosted_engine_setup : Create temporary lock file'} 2022-11-01 21:35:05,260+1100 DEBUG ansible on_any args TASK: ovirt.ovirt.hosted_engine_setup : Create temporary lock file kwargs is_conditional:False 2022-11-01 21:35:05,261+1100 DEBUG ansible on_any args localhost TASK: ovirt.ovirt.hosted_engine_setup : Create temporary lock file kwargs 2022-11-01 21:35:05,889+1100 INFO ansible ok {'status': 'OK', 'ansible_type': 'task', 'ansible_playbook': '/usr/share/ovirt-hosted-engine-setup/he_ansible/trigger_role.yml', 'ansible_host': 'localhost', 'ansible_task': 'Create temporary lock file', 'task_duration': 1} 2022-11-01 21:35:05,890+1100 DEBUG ansible on_any args <ansible.executor.task_result.TaskResult object at 0x7f0b2120eca0> kwargs 2022-11-01 21:35:06,357+1100 DEBUG var changed: host "localhost" var "he_setup_lock_file" type "<class 'dict'>" value: "{ "changed": true, "failed": false, "gid": 0, "group": "root", "mode": "0600", "owner": "root", "path": "/tmp/ansible.volt5pvv_he_setup_lock", "secontext": "unconfined_u:object_r:user_tmp_t:s0", "size": 0, "state": "file", "uid": 0 }" 2022-11-01 21:35:06,358+1100 INFO ansible task start {'status': 'OK', 'ansible_type': 'task', 'ansible_playbook': '/usr/share/ovirt-hosted-engine-setup/he_ansible/trigger_role.yml', 'ansible_task': 'ovirt.ovirt.hosted_engine_setup : Pause execution until /tmp/ansible.volt5pvv_he_setup_lock is removed, delete it once ready to proceed'} 2022-11-01 21:35:06,358+1100 DEBUG ansible on_any args TASK: ovirt.ovirt.hosted_engine_setup : Pause execution until /tmp/ansible.volt5pvv_he_setup_lock is removed, delete it once ready to proceed kwargs is_conditional:False 2022-11-01 21:35:06,359+1100 DEBUG ansible on_any args localhost TASK: ovirt.ovirt.hosted_engine_setup : Pause execution until {{ he_setup_lock_file.path }} is removed, delete it once ready to proceed kwargs ~~~ So from this I can see a couple of things: 1) There's a warning about memory usage of the host (its got 64 GiB BTW). 2) There's a warning that 'ovirtmgmt' is missing from the host (I thought that that network was automatically created). 3) There's an error that: "Host ovirt_node_1.mynet.local installation failed. Failed to configure management network on the host." (this, I assume, is related to the above). 4) There's an error that: "Failed to configure management network on host ovirt_node_1.mynet.local due to setup networks failure." 5) There's an error that vdsm doesn't "like" the bond that I've got set up. I can't work out why, though, because what the script requires is met and exceeded by what is actually there (at least that's how I interpreted it - but, as always, I may be wrong). 6) There's an error that: "Installing Host ovirt_node_1.mynet.local. Check for LVM filter configuration error: Cannot configure LVM filter on host, please run: vdsm-tool config-lvm-filter." Running that command tells me that there are no LVM filters in place - no, I haven't put them in place yet: I'm waiting to hear what is said here. 7) There's an alert that: "Failed to verify Power Management configuration for Host ovirt_node_1.mynet.local." I haven't set up Power Management yet, as I thought that needed to be done from inside the Engine, & I'm not at that stage yet. 8) There's an alert that: "Update to network ovirtmgmt was not applied to virtual network interfaces [<UNKNOWN>]. The actual configuration on the interfaces may differ from the displayed one." Also, I haven't been able to connect to the Web GUI. I suspect that that is some sort of firewall/connection error because I can't ping the Engine from the network, but can ping it from the host (which is headless, so no GUI on it) and I can ping to any node on the network from the engine. Not sure how to fix this one. Also, unless I've got my wires completely crossed, the port that we're suppossed to conect to on the Engine is port 9090 (cockpit) and there's a recent issue (like, literally in the last 15-odd days recent) with cockpit bitching and refusing to work with the latest versions of Firefox & Chrome (https://cockpit-project.org/blog/login-issues.html) - and it looks like there's a fix (I just checked when I grabbed the URL), but is that new version in the (EL8) repos? So, what I'm going to do is strip everything out of the 3 self-hosted nodes (software-wise) and start from scratch - tomorrow (its 21:00 local here @ the moment). In the meantime, if people would be kind enough to comment with fixes/suggestions, I'd appreciate it. :-) Cheers Dulux-Oz

On 2. 11. 2022, at 11:01, Matthew J Black <matthew@peregrineit.net> wrote:
OK, so as I said I was going to do I've now gone through the logs.
I've place the log files into DropBox (https://www.dropbox.com/sh/eymwdy8hzn3sa7z/AACscSP2eaFfoiN-QzyeEVfaa?dl=0)
There was only one significant part of the logs (at least that what it appears to me) and I've included that extract below:
ovirt-hosted-engine-setup-ansible-bootstrap_local_vm-...log Extract
~~~ 2022-11-01 21:34:57,395+1100 INFO ansible task start {'status': 'OK', 'ansible_type': 'task', 'ansible_playbook': '/usr/share/ovirt-hosted-engine-setup/he_ansible/trigger_role.yml', 'ansible_task': 'ovirt.ovirt.hosted_engine_setup : Notify the user about a failure'} 2022-11-01 21:34:57,395+1100 DEBUG ansible on_any args TASK: ovirt.ovirt.hosted_engine_setup : Notify the user about a failure kwargs is_conditional:False 2022-11-01 21:34:57,396+1100 DEBUG ansible on_any args localhost TASK: ovirt.ovirt.hosted_engine_setup : Notify the user about a failure kwargs 2022-11-01 21:34:57,875+1100 INFO ansible skipped {'status': 'SKIPPED', 'ansible_type': 'task', 'ansible_playbook': '/usr/share/ovirt-hosted-engine-setup/he_ansible/trigger_role.yml', 'ansible_task': 'Notify the user about a failure', 'ansible_host': 'localhost'} 2022-11-01 21:34:57,876+1100 DEBUG ansible on_any args <ansible.executor.task_result.TaskResult object at 0x7f0b21370ee0> kwargs 2022-11-01 21:34:58,359+1100 INFO ansible task start {'status': 'OK', 'ansible_type': 'task', 'ansible_playbook': '/usr/share/ovirt-hosted-engine-setup/he_ansible/trigger_role.yml', 'ansible_task': 'ovirt.ovirt.hosted_engine_setup : Set host_id'} 2022-11-01 21:34:58,359+1100 DEBUG ansible on_any args TASK: ovirt.ovirt.hosted_engine_setup : Set host_id kwargs is_conditional:False 2022-11-01 21:34:58,360+1100 DEBUG ansible on_any args localhost TASK: ovirt.ovirt.hosted_engine_setup : Set host_id kwargs 2022-11-01 21:34:58,844+1100 DEBUG var changed: host "localhost" var "host_id" type "<class 'ansible.utils.unsafe_proxy.AnsibleUnsafeText'>" value: ""eb33e62a-2929-499f-80de-b7ac38a075f5"" 2022-11-01 21:34:58,844+1100 INFO ansible ok {'status': 'OK', 'ansible_type': 'task', 'ansible_playbook': '/usr/share/ovirt-hosted-engine-setup/he_ansible/trigger_role.yml', 'ansible_host': 'localhost', 'ansible_task': 'Set host_id', 'task_duration': 0} 2022-11-01 21:34:58,844+1100 DEBUG ansible on_any args <ansible.executor.task_result.TaskResult object at 0x7f0b214c4a00> kwargs 2022-11-01 21:34:59,288+1100 INFO ansible task start {'status': 'OK', 'ansible_type': 'task', 'ansible_playbook': '/usr/share/ovirt-hosted-engine-setup/he_ansible/trigger_role.yml', 'ansible_task': 'ovirt.ovirt.hosted_engine_setup : Collect error events from the Engine'} 2022-11-01 21:34:59,289+1100 DEBUG ansible on_any args TASK: ovirt.ovirt.hosted_engine_setup : Collect error events from the Engine kwargs is_conditional:False 2022-11-01 21:34:59,290+1100 DEBUG ansible on_any args localhost TASK: ovirt.ovirt.hosted_engine_setup : Collect error events from the Engine kwargs 2022-11-01 21:35:00,157+1100 INFO ansible ok {'status': 'OK', 'ansible_type': 'task', 'ansible_playbook': '/usr/share/ovirt-hosted-engine-setup/he_ansible/trigger_role.yml', 'ansible_host': 'localhost', 'ansible_task': 'Collect error events from the Engine', 'task_duration': 1} 2022-11-01 21:35:00,157+1100 DEBUG ansible on_any args <ansible.executor.task_result.TaskResult object at 0x7f0b214ae760> kwargs 2022-11-01 21:35:00,625+1100 DEBUG var changed: host "localhost" var "error_events" type "<class 'dict'>" value: "{ "changed": false, "failed": false, "ovirt_events": [ { "cluster": { "href": "/ovirt-engine/api/clusters/c44e2594-989d-4f1e-8308-feec46918d67", "id": "c44e2594-989d-4f1e-8308-feec46918d67", "name": "my_cluster_1" }, "code": 532, "custom_id": -1, "description": "Used memory of host ovirt_node_1.mynet.local in cluster my_cluster_1 [100%] exceeded defined threshold [95%].", "flood_rate": 0, "host": { "href": "/ovirt-engine/api/hosts/eb33e62a-2929-499f-80de-b7ac38a075f5", "id": "eb33e62a-2929-499f-80de-b7ac38a075f5", "name": "ovirt_node_1.mynet.local" }, "href": "/ovirt-engine/api/events/142", "id": "142", "index": 142, "origin": "oVirt", "severity": "warning", "time": "2022-11-01 21:34:57.640000+11:00" }, { "cluster": { "href": "/ovirt-engine/api/clusters/c44e2594-989d-4f1e-8308-feec46918d67", "id": "c44e2594-989d-4f1e-8308-feec46918d67", "name": "my_cluster_1" }, "code": 519, "correlation_id": "65a04e79", "custom_id": -1, "description": "Host ovirt_node_1.mynet.local does not comply with the cluster my_cluster_1 networks, the following networks are missing on host: 'ovirtmgmt'", "flood_rate": 0, "host": { "href": "/ovirt-engine/api/hosts/eb33e62a-2929-499f-80de-b7ac38a075f5", "id": "eb33e62a-2929-499f-80de-b7ac38a075f5", "name": "ovirt_node_1.mynet.local" }, "href": "/ovirt-engine/api/events/140", "id": "140", "index": 140, "origin": "oVirt", "severity": "warning", "time": "2022-11-01 21:34:57.404000+11:00", "user": { "name": "SYSTEM" } }, { "cluster": { "href": "/ovirt-engine/api/clusters/c44e2594-989d-4f1e-8308-feec46918d67", "id": "c44e2594-989d-4f1e-8308-feec46918d67", "name": "my_cluster_1" }, "code": 505, "correlation_id": "8a77e33b-64f8-409b-abe1-78a37ae6df4c", "custom_id": -1, "description": "Host ovirt_node_1.mynet.local installation failed. Failed to configure management network on the host.", "flood_rate": 0, "host": { "href": "/ovirt-engine/api/hosts/eb33e62a-2929-499f-80de-b7ac38a075f5", "id": "eb33e62a-2929-499f-80de-b7ac38a075f5", "name": "ovirt_node_1.mynet.local" }, "href": "/ovirt-engine/api/events/137", "id": "137", "index": 137, "origin": "oVirt", "severity": "error", "time": "2022-11-01 21:34:51.506000+11:00", "user": { "href": "/ovirt-engine/api/users/66aaaec1-29f1-4e4f-a8c6-2a9c7f3319c2", "id": "66aaaec1-29f1-4e4f-a8c6-2a9c7f3319c2", "name": "admin@ovirt@internalkeycloak-authz" } }, { "code": 1120, "custom_id": -1, "description": "Failed to configure management network on host ovirt_node_1.mynet.local due to setup networks failure.", "flood_rate": 0, "host": { "href": "/ovirt-engine/api/hosts/eb33e62a-2929-499f-80de-b7ac38a075f5", "id": "eb33e62a-2929-499f-80de-b7ac38a075f5", "name": "ovirt_node_1.mynet.local" }, "href": "/ovirt-engine/api/events/136", "id": "136", "index": 136, "origin": "oVirt", "severity": "error", "time": "2022-11-01 21:34:51.413000+11:00" }, { "code": 10802, "custom_id": -1, "description": "VDSM ovirt_node_1.mynet.local command HostSetupNetworksVDS failed: Internal JSON-RPC error: {'reason': \"\\ndesired\\n=======\\n---\\nname: bond_1\\ntype: bond\\nstate: up\\nipv4:\\n enabled: false\\nipv6:\\n enabled: false\\nlink-aggregation:\\n mode: 802.3ad\\n options:\\n downdelay: 1000\\n miimon: 1000\\n updelay: 1000\\n port:\\n - eno1\\nmac-address: 3C:EC:EF:83:77:4C\\nmtu: 1500\\n\\ncurrent\\n=======\\n---\\nname: bond_1\\ntype: bond\\nstate: up\\naccept-all-mac-addresses: false\\nethtool:\\n feature:\\n esp-hw-offload: false\\n esp-tx-csum-hw-offload: false\\n highdma: true\\n rx-gro: true\\n rx-gro-list: false\\n rx-lro: false\\n rx-udp-gro-forwarding: false\\n rx-vlan-filter: true\\n rx-vlan-hw-parse: true\\n tx-checksum-ip-generic: true\\n tx-esp-segmentation: false\\n tx-generic-segmentation: true\\n tx-gre-csum-segmentation: true\\n tx-gre-segmentation: true\\n tx-gso-list: false\\n tx-ipxip4-segmentation: true\\n tx-ipxip6-segmentation: true\\n tx-nocache-copy: false\\n tx-scatter-gather-fraglist: false\\n tx-sctp-segmentation: false\\n tx-tcp-ecn-segmentation: true\\n tx-tcp-mangleid-segmentation: true\\n tx-tcp-segmentation: true\\n tx-tcp6-segmentation: true\\n tx-udp-segmentation: true\\n tx-udp_tnl-csum-segmentation: true\\n tx-udp_tnl-segmentation: true\\nipv4:\\n enabled: false\\nipv6:\\n enabled: false\\nlink-aggregation:\\n mode: 802.3ad\\n options:\\n ad_actor_sys_prio: 65535\\n ad_actor_system: 00:00:00:00:00:00\\n ad_select: stable\\n ad_user_port_key: 0\\n all_slaves_active: dropped\\n arp_all_targets: any\\n arp_interval: 0\\n arp_ip_target: ''\\n arp_validate: none\\n downdelay: 1000\\n lacp_rate: slow\\n miimon: 1000\\n min_links: 0\\n updelay: 1000\\n use_carrier: true\\n xmit_hash_policy: layer2\\n port:\\n - eno1\\n - eno2\\nlldp:\\n enabled: false\\ nmac-address: 3C:EC:EF:83:77:4C\\nmtu: 1500\\n\\ndifference\\n==========\\n--- desired\\n+++ current\\n@@ -2,6 +2,36 @@\\n name: bond_1\\n type: bond\\n state: up\\n+accept-all-mac-addresses: false\\n+ethtool:\\n+ feature:\\n+ esp-hw-offload: false\\n+ esp-tx-csum-hw-offload: false\\n+ highdma: true\\n+ rx-gro: true\\n+ rx-gro-list: false\\n+ rx-lro: false\\n+ rx-udp-gro-forwarding: false\\n+ rx-vlan-filter: true\\n+ rx-vlan-hw-parse: true\\n+ tx-checksum-ip-generic: true\\n+ tx-esp-segmentation: false\\n+ tx-generic-segmentation: true\\n+ tx-gre-csum-segmentation: true\\n+ tx-gre-segmentation: true\\n+ tx-gso-list: false\\n+ tx-ipxip4-segmentation: true\\n+ tx-ipxip6-segmentation: true\\n+ tx-nocache-copy: false\\n+ tx-scatter-gather-fraglist: false\\n+ tx-sctp-segmentation: false\\n+ tx-tcp-ecn-segmentation: true\\n+ tx-tcp-mangleid-segmentation: true\\n+ tx-tcp-segmentation: true\\n+ tx-tcp6-segmentation: tr ue\\n+ tx-udp-segmentation: true\\n+ tx-udp_tnl-csum-segmentation: true\\n+ tx-udp_tnl-segmentation: true\\n ipv4:\\n enabled: false\\n ipv6:\\n@@ -9,10 +39,26 @@\\n link-aggregation:\\n mode: 802.3ad\\n options:\\n+ ad_actor_sys_prio: 65535\\n+ ad_actor_system: 00:00:00:00:00:00\\n+ ad_select: stable\\n+ ad_user_port_key: 0\\n+ all_slaves_active: dropped\\n+ arp_all_targets: any\\n+ arp_interval: 0\\n+ arp_ip_target: ''\\n+ arp_validate: none\\n downdelay: 1000\\n+ lacp_rate: slow\\n miimon: 1000\\n+ min_links: 0\\n updelay: 1000\\n+ use_carrier: true\\n+ xmit_hash_policy: layer2\\n port:\\n - eno1\\n+ - eno2\\n+lldp:\\n+ enabled: false\\n mac-address: 3C:EC:EF:83:77:4C\\n mtu: 1500\\n\\n\"}", "flood_rate": 0, "host": { "href": "/ovirt-engine/api/hosts/eb33e62a-2929-499f-80de-b7ac38a075f5", "id": "eb33e62a-2929-499f-80de-b7ac38a075f5", "name": "ovirt_node_1.mynet.local" }, "href": "/ovirt-engine/api/events/135", "id": "135", "index": 135, "origin": "oVirt", "severity": "error", "time": "2022-11-01 21:34:51.387000+11:00" }, { "cluster": { "href": "/ovirt-engine/api/clusters/c44e2594-989d-4f1e-8308-feec46918d67", "id": "c44e2594-989d-4f1e-8308-feec46918d67", "name": "my_cluster_1" }, "code": 553, "correlation_id": "8a77e33b-64f8-409b-abe1-78a37ae6df4c", "custom_id": -1, "description": "Installing Host ovirt_node_1.mynet.local. Check for LVM filter configuration error: Cannot configure LVM filter on host, please run: vdsm-tool config-lvm-filter.", "flood_rate": 0, "host": { "href": "/ovirt-engine/api/hosts/eb33e62a-2929-499f-80de-b7ac38a075f5", "id": "eb33e62a-2929-499f-80de-b7ac38a075f5", "name": "ovirt_node_1.mynet.local" }, "href": "/ovirt-engine/api/events/59", "id": "59", "index": 59, "origin": "oVirt", "severity": "error", "time": "2022-11-01 21:32:19.692000+11:00" }, { "code": 9000, "custom_id": -1, "description": "Failed to verify Power Management configuration for Host ovirt_node_1.mynet.local.", "flood_rate": 0, "host": { "href": "/ovirt-engine/api/hosts/eb33e62a-2929-499f-80de-b7ac38a075f5", "id": "eb33e62a-2929-499f-80de-b7ac38a075f5", "name": "ovirt_node_1.mynet.local" }, "href": "/ovirt-engine/api/events/13", "id": "13", "index": 13, "origin": "oVirt", "severity": "alert", "time": "2022-11-01 21:31:44.645000+11:00" }, { "code": 11291, "custom_id": -1, "description": "Update to network ovirtmgmt was not applied to virtual network interfaces [<UNKNOWN>]. The actual configuration on the interfaces may differ from the displayed one.", "flood_rate": 0, "href": "/ovirt-engine/api/events/10", "id": "10", "index": 10, "origin": "oVirt", "severity": "alert", "time": "2022-11-01 21:31:35.606000+11:00" } ] }" 2022-11-01 21:35:00,625+1100 INFO ansible task start {'status': 'OK', 'ansible_type': 'task', 'ansible_playbook': '/usr/share/ovirt-hosted-engine-setup/he_ansible/trigger_role.yml', 'ansible_task': 'ovirt.ovirt.hosted_engine_setup : Generate the error message from the engine events'} 2022-11-01 21:35:00,625+1100 DEBUG ansible on_any args TASK: ovirt.ovirt.hosted_engine_setup : Generate the error message from the engine events kwargs is_conditional:False 2022-11-01 21:35:00,626+1100 DEBUG ansible on_any args localhost TASK: ovirt.ovirt.hosted_engine_setup : Generate the error message from the engine events kwargs 2022-11-01 21:35:01,125+1100 DEBUG var changed: host "localhost" var "error_description" type "<class 'ansible.utils.unsafe_proxy.AnsibleUnsafeText'>" value: "" code 505: Host ovirt_node_1.mynet.local installation failed. Failed to configure management network on the host., code 519: Host ovirt_node_1.mynet.local does not comply with the cluster my_cluster_1 networks, the following networks are missing on host: 'ovirtmgmt', code 532: Used memory of host ovirt_node_1.mynet.local in cluster my_cluster_1 [100%] exceeded defined threshold [95%]., code 553: Installing Host ovirt_node_1.mynet.local. Check for LVM filter configuration error: Cannot configure LVM filter on host, please run: vdsm-tool config-lvm-filter., code 1120: Failed to configure management network on host ovirt_node_1.mynet.local due to setup networks failure., code 9000: Failed to verify Power Management configuration for Host ovirt_node_1.mynet.local., code 10802: VDSM ovirt_node_1.mynet.local comma nd HostSetupNetworksVDS failed: Internal JSON-RPC error: {'reason': \"\\ndesired\\n=======\\n---\\nname: bond_1\\ntype: bond\\nstate: up\\nipv4:\\n enabled: false\\nipv6:\\n enabled: false\\nlink-aggregation:\\n mode: 802.3ad\\n options:\\n downdelay: 1000\\n miimon: 1000\\n updelay: 1000\\n port:\\n - eno1\\nmac-address: 3C:EC:EF:83:77:4C\\nmtu: 1500\\n\\ncurrent\\n=======\\n---\\nname: bond_1\\ntype: bond\\nstate: up\\naccept-all-mac-addresses: false\\nethtool:\\n feature:\\n esp-hw-offload: false\\n esp-tx-csum-hw-offload: false\\n highdma: true\\n rx-gro: true\\n rx-gro-list: false\\n rx-lro: false\\n rx-udp-gro-forwarding: false\\n rx-vlan-filter: true\\n rx-vlan-hw-parse: true\\n tx-checksum-ip-generic: true\\n tx-esp-segmentation: false\\n tx-generic-segmentation: true\\n tx-gre-csum-segmentation: true\\n tx-gre-segmentation: true\\n tx-gso-list: false\\n tx-ipxip4-segmentation: true\\n tx-ipxip6-segmentation: true\\n tx-nocache-copy: false\\n tx-scatter-gather-fraglist: false\\n tx-sctp-segmentation: false\\n tx-tcp-ecn-segmentation: true\\n tx-tcp-mangleid-segmentation: true\\n tx-tcp-segmentation: true\\n tx-tcp6-segmentation: true\\n tx-udp-segmentation: true\\n tx-udp_tnl-csum-segmentation: true\\n tx-udp_tnl-segmentation: true\\nipv4:\\n enabled: false\\nipv6:\\n enabled: false\\nlink-aggregation:\\n mode: 802.3ad\\n options:\\n ad_actor_sys_prio: 65535\\n ad_actor_system: 00:00:00:00:00:00\\n ad_select: stable\\n ad_user_port_key: 0\\n all_slaves_active: dropped\\n arp_all_targets: any\\n arp_interval: 0\\n arp_ip_target: ''\\n arp_validate: none\\n downdelay: 1000\\n lacp_rate: slow\\n miimon: 1000\\n min_links: 0\\n updelay: 1000\\n use_carrier: true\\n xmit_hash_policy: layer2\\n port:\\n - eno1\\n - eno2\\nlldp:\\n enabled: false\\nmac-address: 3C:EC:EF:83:77:4C\\nmtu: 1500\\n\\ndifference\\n ==========\\n--- desired\\n+++ current\\n@@ -2,6 +2,36 @@\\n name: bond_1\\n type: bond\\n state: up\\n+accept-all-mac-addresses: false\\n+ethtool:\\n+ feature:\\n+ esp-hw-offload: false\\n+ esp-tx-csum-hw-offload: false\\n+ highdma: true\\n+ rx-gro: true\\n+ rx-gro-list: false\\n+ rx-lro: false\\n+ rx-udp-gro-forwarding: false\\n+ rx-vlan-filter: true\\n+ rx-vlan-hw-parse: true\\n+ tx-checksum-ip-generic: true\\n+ tx-esp-segmentation: false\\n+ tx-generic-segmentation: true\\n+ tx-gre-csum-segmentation: true\\n+ tx-gre-segmentation: true\\n+ tx-gso-list: false\\n+ tx-ipxip4-segmentation: true\\n+ tx-ipxip6-segmentation: true\\n+ tx-nocache-copy: false\\n+ tx-scatter-gather-fraglist: false\\n+ tx-sctp-segmentation: false\\n+ tx-tcp-ecn-segmentation: true\\n+ tx-tcp-mangleid-segmentation: true\\n+ tx-tcp-segmentation: true\\n+ tx-tcp6-segmentation: true\\n+ tx-udp-segmentation: true\\n+ tx-udp_tnl-csum-seg mentation: true\\n+ tx-udp_tnl-segmentation: true\\n ipv4:\\n enabled: false\\n ipv6:\\n@@ -9,10 +39,26 @@\\n link-aggregation:\\n mode: 802.3ad\\n options:\\n+ ad_actor_sys_prio: 65535\\n+ ad_actor_system: 00:00:00:00:00:00\\n+ ad_select: stable\\n+ ad_user_port_key: 0\\n+ all_slaves_active: dropped\\n+ arp_all_targets: any\\n+ arp_interval: 0\\n+ arp_ip_target: ''\\n+ arp_validate: none\\n downdelay: 1000\\n+ lacp_rate: slow\\n miimon: 1000\\n+ min_links: 0\\n updelay: 1000\\n+ use_carrier: true\\n+ xmit_hash_policy: layer2\\n port:\\n - eno1\\n+ - eno2\\n+lldp:\\n+ enabled: false\\n mac-address: 3C:EC:EF:83:77:4C\\n mtu: 1500\\n\\n\"}, "" 2022-11-01 21:35:01,125+1100 INFO ansible ok {'status': 'OK', 'ansible_type': 'task', 'ansible_playbook': '/usr/share/ovirt-hosted-engine-setup/he_ansible/trigger_role.yml', 'ansible_host': 'localhost', 'ansible_task': 'Generate the error message from the engine events', 'task_duration': 0} 2022-11-01 21:35:01,125+1100 DEBUG ansible on_any args <ansible.executor.task_result.TaskResult object at 0x7f0b21370ee0> kwargs 2022-11-01 21:35:01,583+1100 INFO ansible task start {'status': 'OK', 'ansible_type': 'task', 'ansible_playbook': '/usr/share/ovirt-hosted-engine-setup/he_ansible/trigger_role.yml', 'ansible_task': 'ovirt.ovirt.hosted_engine_setup : Notify with error description'} 2022-11-01 21:35:01,583+1100 DEBUG ansible on_any args TASK: ovirt.ovirt.hosted_engine_setup : Notify with error description kwargs is_conditional:False 2022-11-01 21:35:01,584+1100 DEBUG ansible on_any args localhost TASK: ovirt.ovirt.hosted_engine_setup : Notify with error description kwargs 2022-11-01 21:35:02,032+1100 INFO ansible ok {'status': 'OK', 'ansible_type': 'task', 'ansible_playbook': '/usr/share/ovirt-hosted-engine-setup/he_ansible/trigger_role.yml', 'ansible_host': 'localhost', 'ansible_task': 'Notify with error description', 'task_duration': 0} 2022-11-01 21:35:02,032+1100 DEBUG ansible on_any args <ansible.executor.task_result.TaskResult object at 0x7f0b214ae790> kwargs 2022-11-01 21:35:02,511+1100 INFO ansible task start {'status': 'OK', 'ansible_type': 'task', 'ansible_playbook': '/usr/share/ovirt-hosted-engine-setup/he_ansible/trigger_role.yml', 'ansible_task': 'ovirt.ovirt.hosted_engine_setup : Notify with generic error'} 2022-11-01 21:35:02,511+1100 DEBUG ansible on_any args TASK: ovirt.ovirt.hosted_engine_setup : Notify with generic error kwargs is_conditional:False 2022-11-01 21:35:02,512+1100 DEBUG ansible on_any args localhost TASK: ovirt.ovirt.hosted_engine_setup : Notify with generic error kwargs 2022-11-01 21:35:02,954+1100 INFO ansible skipped {'status': 'SKIPPED', 'ansible_type': 'task', 'ansible_playbook': '/usr/share/ovirt-hosted-engine-setup/he_ansible/trigger_role.yml', 'ansible_task': 'Notify with generic error', 'ansible_host': 'localhost'} 2022-11-01 21:35:02,954+1100 DEBUG ansible on_any args <ansible.executor.task_result.TaskResult object at 0x7f0b21077520> kwargs 2022-11-01 21:35:03,401+1100 INFO ansible task start {'status': 'OK', 'ansible_type': 'task', 'ansible_playbook': '/usr/share/ovirt-hosted-engine-setup/he_ansible/trigger_role.yml', 'ansible_task': 'ovirt.ovirt.hosted_engine_setup : Let the user connect to the bootstrap engine to manually fix host configuration'} 2022-11-01 21:35:03,402+1100 DEBUG ansible on_any args TASK: ovirt.ovirt.hosted_engine_setup : Let the user connect to the bootstrap engine to manually fix host configuration kwargs is_conditional:False 2022-11-01 21:35:03,402+1100 DEBUG ansible on_any args localhost TASK: ovirt.ovirt.hosted_engine_setup : Let the user connect to the bootstrap engine to manually fix host configuration kwargs 2022-11-01 21:35:03,843+1100 INFO ansible ok {'status': 'OK', 'ansible_type': 'task', 'ansible_playbook': '/usr/share/ovirt-hosted-engine-setup/he_ansible/trigger_role.yml', 'ansible_host': 'localhost', 'ansible_task': 'Let the user connect to the bootstrap engine to manually fix host configuration', 'task_duration': 0} 2022-11-01 21:35:03,844+1100 DEBUG ansible on_any args <ansible.executor.task_result.TaskResult object at 0x7f0b214ae550> kwargs 2022-11-01 21:35:04,322+1100 INFO ansible task start {'status': 'OK', 'ansible_type': 'task', 'ansible_playbook': '/usr/share/ovirt-hosted-engine-setup/he_ansible/trigger_role.yml', 'ansible_task': 'ovirt.ovirt.hosted_engine_setup : include_tasks'} 2022-11-01 21:35:04,323+1100 DEBUG ansible on_any args TASK: ovirt.ovirt.hosted_engine_setup : include_tasks kwargs is_conditional:False 2022-11-01 21:35:04,323+1100 DEBUG ansible on_any args localhost TASK: ovirt.ovirt.hosted_engine_setup : include_tasks kwargs 2022-11-01 21:35:04,772+1100 INFO ansible ok {'status': 'OK', 'ansible_type': 'task', 'ansible_playbook': '/usr/share/ovirt-hosted-engine-setup/he_ansible/trigger_role.yml', 'ansible_host': 'localhost', 'ansible_task': '', 'task_duration': 0} 2022-11-01 21:35:04,772+1100 DEBUG ansible on_any args <ansible.executor.task_result.TaskResult object at 0x7f0b2154e700> kwargs 2022-11-01 21:35:04,798+1100 DEBUG ansible on_any args /usr/share/ansible/collections/ansible_collections/ovirt/ovirt/roles/hosted_engine_setup/tasks/pause_execution.yml (args={} vars={}): [localhost] kwargs 2022-11-01 21:35:05,260+1100 INFO ansible task start {'status': 'OK', 'ansible_type': 'task', 'ansible_playbook': '/usr/share/ovirt-hosted-engine-setup/he_ansible/trigger_role.yml', 'ansible_task': 'ovirt.ovirt.hosted_engine_setup : Create temporary lock file'} 2022-11-01 21:35:05,260+1100 DEBUG ansible on_any args TASK: ovirt.ovirt.hosted_engine_setup : Create temporary lock file kwargs is_conditional:False 2022-11-01 21:35:05,261+1100 DEBUG ansible on_any args localhost TASK: ovirt.ovirt.hosted_engine_setup : Create temporary lock file kwargs 2022-11-01 21:35:05,889+1100 INFO ansible ok {'status': 'OK', 'ansible_type': 'task', 'ansible_playbook': '/usr/share/ovirt-hosted-engine-setup/he_ansible/trigger_role.yml', 'ansible_host': 'localhost', 'ansible_task': 'Create temporary lock file', 'task_duration': 1} 2022-11-01 21:35:05,890+1100 DEBUG ansible on_any args <ansible.executor.task_result.TaskResult object at 0x7f0b2120eca0> kwargs 2022-11-01 21:35:06,357+1100 DEBUG var changed: host "localhost" var "he_setup_lock_file" type "<class 'dict'>" value: "{ "changed": true, "failed": false, "gid": 0, "group": "root", "mode": "0600", "owner": "root", "path": "/tmp/ansible.volt5pvv_he_setup_lock", "secontext": "unconfined_u:object_r:user_tmp_t:s0", "size": 0, "state": "file", "uid": 0 }" 2022-11-01 21:35:06,358+1100 INFO ansible task start {'status': 'OK', 'ansible_type': 'task', 'ansible_playbook': '/usr/share/ovirt-hosted-engine-setup/he_ansible/trigger_role.yml', 'ansible_task': 'ovirt.ovirt.hosted_engine_setup : Pause execution until /tmp/ansible.volt5pvv_he_setup_lock is removed, delete it once ready to proceed'} 2022-11-01 21:35:06,358+1100 DEBUG ansible on_any args TASK: ovirt.ovirt.hosted_engine_setup : Pause execution until /tmp/ansible.volt5pvv_he_setup_lock is removed, delete it once ready to proceed kwargs is_conditional:False 2022-11-01 21:35:06,359+1100 DEBUG ansible on_any args localhost TASK: ovirt.ovirt.hosted_engine_setup : Pause execution until {{ he_setup_lock_file.path }} is removed, delete it once ready to proceed kwargs ~~~
So from this I can see a couple of things:
1) There's a warning about memory usage of the host (its got 64 GiB BTW).
probably a first bogus report, not relevant
2) There's a warning that 'ovirtmgmt' is missing from the host (I thought that that network was automatically created).
that's what it is trying to do, and failed. That would be the reason for you host being non operational
3) There's an error that: "Host ovirt_node_1.mynet.local installation failed. Failed to configure management network on the host." (this, I assume, is related to the above).
right
4) There's an error that: "Failed to configure management network on host ovirt_node_1.mynet.local due to setup networks failure."
same thing
5) There's an error that vdsm doesn't "like" the bond that I've got set up. I can't work out why, though, because what the script requires is met and exceeded by what is actually there (at least that's how I interpreted it - but, as always, I may be wrong).
it's nmstate unable to reconcile the requested configuration. That is usually caused by some uncommon configuration it can't really deal with. What bond did you create? Do you need it for initial access? In general all network configuration is to be done after installation
6) There's an error that: "Installing Host ovirt_node_1.mynet.local. Check for LVM filter configuration error: Cannot configure LVM filter on host, please run: vdsm-tool config-lvm-filter." Running that command tells me that there are no LVM filters in place - no, I haven't put them in place yet: I'm waiting to hear what is said here.
irrelevant check at this stage
7) There's an alert that: "Failed to verify Power Management configuration for Host ovirt_node_1.mynet.local." I haven't set up Power Management yet, as I thought that needed to be done from inside the Engine, & I'm not at that stage yet.
just a general warning whenever it's not set up to warn that it's not set up:)
8) There's an alert that: "Update to network ovirtmgmt was not applied to virtual network interfaces [<UNKNOWN>]. The actual configuration on the interfaces may differ from the displayed one."
Also, I haven't been able to connect to the Web GUI. I suspect that that is some sort of firewall/connection error because I can't ping the Engine from the network, but can ping it from the host (which is headless, so no GUI on it) and I can ping to any node on the network from the engine. Not sure how to fix this one.
Also, unless I've got my wires completely crossed, the port that we're suppossed to conect to on the Engine is port 9090 (cockpit) and there's a recent issue (like, literally in the last 15-odd days recent) with cockpit bitching and refusing to work with the latest versions of Firefox & Chrome (https://cockpit-project.org/blog/login-issues.html) - and it looks like there's a fix (I just checked when I grabbed the URL), but is that new version in the (EL8) repos?
avoid cockpit altogether. GUI is runnign on standard 80/443 ports
So, what I'm going to do is strip everything out of the 3 self-hosted nodes (software-wise) and start from scratch - tomorrow (its 21:00 local here @ the moment). In the meantime, if people would be kind enough to comment with fixes/suggestions, I'd appreciate it. :-)
I'd try to avoid altering network configuration prior to installation as much as possible Thanks, michal
Cheers
Dulux-Oz _______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-leave@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/GVUOLBFLSBCKJV...

***WOOHOO*** Finally - *FINALLY* - I've got a clean install! As I said in my last post, I stripped the three hosts back to bare metal and reinstalled *everything* from scratch, as per the oVirt doco - and late this afternoon the deployment script finished without any errors and I could pull up the Engine Web page. So all good (except see below) :-) Questions: - When logging in to the Engine URL for the first time, is the username "admin" or "admin@internal"? - Just to make it 100% clear in my own muddled mind: Step 5.4 (of the oVirt CLI Install doco) is performed on the Engine VM, isn't it, and *not* on the host, right? I'm having trouble logging in - but that's OK because I'm pretty sure I know what the issue is (bad password :-) ) so I'm going to tackle that one tomorrow. Thanks for the help everyone (include the OOB help) - if someone could please confirm these final(?) two questions that'd be great. Cheers Dulux-Oz

Actually, I just thought of another question: I've got console root access to the Engine VM: How hard is it to change the URL GUI "admin@local" password from there - and what is the process? I ask because this'll mean I won't have to do a re-install to reset the bad password I mentioned above. Cheers Dulux-Oz

Hi Matthew, It should be as simple as: # ovirt-aaa-jdbc-tool user password-reset admin On Mon, Nov 7, 2022 at 10:27 AM Matthew J Black <matthew@peregrineit.net> wrote:
Actually, I just thought of another question:
I've got console root access to the Engine VM: How hard is it to change the URL GUI "admin@local" password from there - and what is the process?
I ask because this'll mean I won't have to do a re-install to reset the bad password I mentioned above.
Cheers
Dulux-Oz _______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-leave@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/O2F6MYURHW2EHB...
-- Lev Veyde Senior Software Engineer, RHCE | RHCVA | MCITP Red Hat Israel <https://www.redhat.com> lev@redhat.com | lveyde@redhat.com <https://red.ht/sig> TRIED. TESTED. TRUSTED. <https://redhat.com/trusted>

Hi Matthew Sorry if you already solved the problem. The default HE admin name looks like changed to "admin@ovirt". BTW how did you solve the original problem? I also had same problem like "Host is not up..." and I modified the ansible role which calls "ovirt.ovirt.ovirt_host_info" to be skipped.. in the end. Bad solution. Regards, Shin

Hi Shin, Yeah, all sorted, all working - even the stuff that I had crop-up since then. As it stands at the moment we've got a 3-node self-hosted oVirt Cluster using Gluster and iSCSI Storage provided by a hyper-converged Gluster Cluster (the same physical boxes ad the oVirt Hosts) and a 3-node Ceph Cluster using 3 iSCSI Gateways - a total of 6 servers if anyone is counting. There's about 140 TB of storage space, about 187 GB of VM-usable RAM, and about 190 VM-usable GFlops. Cheers Dulux-Oz

Hello Dulux-Oz Thanks for your rep, good to know working fine. Then, what was the problem of the "Host is not up..." issue and how did *you* solve that..? Hope my question not bother you. Regards, Shin

Sorry Shin, I meant to get back to answer this when you asked, but got distracted and then forgot about it (which was rude of me - sorry). I also needed to "document it for myself" because, over a year later, I've ran into the same issue (see https://lists.ovirt.org/archives/list/users@ovirt.org/thread/7CXIPLSXOGL3EDW...). In short, (to now answer your Q and to document it for me, plus anyone else who runs across this) the solution (for me) was to accept the default names for the Data Center and the Cluster and then edit the names to what was required/desired once the engine was successfully deployed. This thread (https://lists.ovirt.org/archives/list/users@ovirt.org/thread/IDLGSBQFX35EHHG...) gave me the clue (the 2nd time around). Cheers Dulux-Oz
participants (8)
-
ajude.pereira@gmail.com
-
kura_amp@hotmail.com
-
Lev Veyde
-
Matthew J Black
-
Matthew J Black
-
Michal Skrivanek
-
Volenbovskyi, Konstantin
-
Yedidyah Bar David