deploy oVirt 4.4 errors

grep ERROR /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20210227232700-gil2fj.log ;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 9786 2021-02-28 00:00:12,059+0600 ERROR otopi.ovirt_hosted_engine_setup.ansible_utils ansible_utils._process_output:109 ovirtsdk4.ConnectionError: Error while sending HTTP request: (7, 'Failed to connect to ovirt4-adm.domain.local port 443: No route to host') 2021-02-28 00:00:12,160+0600 ERROR otopi.ovirt_hosted_engine_setup.ansible_utils ansible_utils._process_output:109 fatal: [localhost]: FAILED! => {"attempts": 50, "changed": false, "msg": "Error while sending HTTP request: (7, 'Failed to connect to ovirt4-adm.domain.local port 443: No route to host')"} 2021-02-28 00:00:58,055+0600 ERROR otopi.ovirt_hosted_engine_setup.ansible_utils ansible_utils._process_output:109 fatal: [localhost]: FAILED! => {"changed": false, "msg": "The system may not be provisioned according to the playbook results: please check the logs for the issue, fix accordingly or re-deploy from scratch.\n"} 2021-02-28 00:00:58,759+0600 ERROR otopi.context context._executeMethod:154 Failed to execute stage 'Closing up': Failed executing ansible-playbook 2021-02-28 00:01:05,984+0600 ERROR otopi.ovirt_hosted_engine_setup.ansible_utils ansible_utils._process_output:109 fatal: [localhost]: UNREACHABLE! => {"changed": false, "msg": "Failed to connect to the host via ssh: ssh: connect to host ovirt4-adm.domain.local. port 22: No route to host", "skip_reason": "Host localhost is unreachable", "unreachable": true} 2021-02-28 00:01:22,146+0600 ERROR otopi.plugins.gr_he_common.core.misc misc._terminate:167 Hosted Engine deployment failed: please check the logs for the issue, fix accordingly or re-deploy from scratch.

hosted-engine --deploy script create default interface virbr0 (192.168.1.1/24) instead of ovirtmgmt with static ip in domain.local which includes slave eno1. how to make the bridge right?

On Mon, Mar 1, 2021 at 10:45 AM <grig.4n@gmail.com> wrote:
hosted-engine --deploy script create default interface virbr0 (192.168.1.1/24)
This is a normal middle state of a node-zero hosted-engine deployment. See also, if interested: https://www.ovirt.org/images/Hosted-Engine-4.3-deep-dive.pdf
instead of ovirtmgmt with static ip in domain.local which includes slave eno1. how to make the bridge right?
You should make sure name resolution works well. The deploy process can partially help you with this, see the question about updating /etc/hosts. Good luck and best regards, -- Didi

node02.domain.local resolves normally from domain.local and locahostl. Do I need to create bridge ovirtmgmt with eno1 before hosted-engine --deploy? or create bridge between eno1 and virbr0 to make the ovirt engine available from domain.local?

On Mon, Mar 1, 2021 at 1:50 PM <grig.4n@gmail.com> wrote:
node02.domain.local resolves normally from domain.local and locahostl. Do I need to create bridge ovirtmgmt with eno1 before hosted-engine --deploy? or create bridge between eno1 and virbr0 to make the ovirt engine available from domain.local?
No, you should let 'hosted-engine --deploy' create it for you, on the interface you supply.
_______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-leave@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/QVPAGR5IB24XDO...
-- Didi

but it didn't work. hosted-engine --deploy created nat bridge by default 192.168.1.1/24, ovirt-engine uses ip from dhcp and is not accessible from domain.local

On Mon, Mar 1, 2021 at 2:31 PM <grig.4n@gmail.com> wrote:
but it didn't work. hosted-engine --deploy created nat bridge by default 192.168.1.1/24, ovirt-engine uses ip from dhcp and is not accessible from domain.local
Please check/share all relevant logs: /var/log/ovirt-hosted-engine*/* (including subdirs) and /var/log/vdsm/* from the host, and /var/log/ovirt-engine/* from the engine vm (if you can access it). Also please clarify relevant details about your env - names, name resolution (dns or /etc/hosts), etc. Thanks and best regards,
_______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-leave@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/BEHORVJ3CNYHPV...
-- Didi

ovirt-hosted-engine-cleanup virsh net-destroy default virsh net-undefine default cat /etc/hosts 127.0.0.1 localhost localhost.localdomain localhost4 localhost4.localdomain4 ::1 localhost localhost.localdomain localhost6 localhost6.localdomain6 10.11.120.142 node02.domain.local 10.5.128.142 node02.servers.net 10.11.120.144 ovirt4-adm.domain.local ip a l 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000 link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 inet 127.0.0.1/8 scope host lo valid_lft forever preferred_lft forever inet6 ::1/128 scope host valid_lft forever preferred_lft forever 2: eno3: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500 qdisc mq state DOWN group default qlen 1000 link/ether ec:f4:bb:cb:44:e4 brd ff:ff:ff:ff:ff:ff 3: eno4: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP group default qlen 1000 link/ether ec:f4:bb:cb:44:e5 brd ff:ff:ff:ff:ff:ff 4: eno1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP group default qlen 1000 link/ether ec:f4:bb:cb:44:e0 brd ff:ff:ff:ff:ff:ff inet 10.11.120.142/22 brd 10.11.123.255 scope global noprefixroute eno1 valid_lft forever preferred_lft forever inet6 fe80::eef4:bbff:fecb:44e0/64 scope link valid_lft forever preferred_lft forever 5: eno2: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP group default qlen 1000 link/ether ec:f4:bb:cb:44:e2 brd ff:ff:ff:ff:ff:ff inet 10.5.128.142/24 brd 10.5.128.255 scope global noprefixroute eno2 valid_lft forever preferred_lft forever inet6 fe80::eef4:bbff:fecb:44e2/64 scope link valid_lft forever preferred_lft forever do I need to remove something else before hosted-engine --deploy?

On Mon, Mar 1, 2021 at 3:31 PM <grig.4n@gmail.com> wrote:
ovirt-hosted-engine-cleanup virsh net-destroy default virsh net-undefine default cat /etc/hosts 127.0.0.1 localhost localhost.localdomain localhost4 localhost4.localdomain4 ::1 localhost localhost.localdomain localhost6 localhost6.localdomain6 10.11.120.142 node02.domain.local 10.5.128.142 node02.servers.net 10.11.120.144 ovirt4-adm.domain.local ip a l 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000 link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 inet 127.0.0.1/8 scope host lo valid_lft forever preferred_lft forever inet6 ::1/128 scope host valid_lft forever preferred_lft forever 2: eno3: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500 qdisc mq state DOWN group default qlen 1000 link/ether ec:f4:bb:cb:44:e4 brd ff:ff:ff:ff:ff:ff 3: eno4: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP group default qlen 1000 link/ether ec:f4:bb:cb:44:e5 brd ff:ff:ff:ff:ff:ff 4: eno1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP group default qlen 1000 link/ether ec:f4:bb:cb:44:e0 brd ff:ff:ff:ff:ff:ff inet 10.11.120.142/22 brd 10.11.123.255 scope global noprefixroute eno1 valid_lft forever preferred_lft forever inet6 fe80::eef4:bbff:fecb:44e0/64 scope link valid_lft forever preferred_lft forever 5: eno2: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP group default qlen 1000 link/ether ec:f4:bb:cb:44:e2 brd ff:ff:ff:ff:ff:ff inet 10.5.128.142/24 brd 10.5.128.255 scope global noprefixroute eno2 valid_lft forever preferred_lft forever inet6 fe80::eef4:bbff:fecb:44e2/64 scope link valid_lft forever preferred_lft forever
do I need to remove something else before hosted-engine --deploy?
Looks ok to me. Please note that 'ovirt-hosted-engine-cleanup' is not perfect - if in doubt, better reinstall the OS. -- Didi

strangely, left the domain and cluster name default. ovirt engine is available at node02.domain.local:6900/ovirt-engine. but it has no information about node2: cpu, networks, ram. hosted-engine --deploy stopped at "TASK [ovirt.ovirt.hosted_engine_setup: Wait for host to be up]" -- С уважением, Григорий Ч

but the same errors over time grep ERROR /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20210301204742-wmcz7j.log ;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 24809 2021-03-01 21:34:44,426+0600 ERROR otopi.ovirt_hosted_engine_setup.ansible_utils ansible_utils._process_output:109 fatal: [localhost]: FAILED! => {"changed": false, "msg": "Host is not up, please check logs, perhaps also on the engine machine"} 2021-03-01 21:35:35,829+0600 ERROR otopi.ovirt_hosted_engine_setup.ansible_utils ansible_utils._process_output:109 fatal: [localhost]: FAILED! => {"changed": false, "msg": "The system may not be provisioned according to the playbook results: please check the logs for the issue, fix accordingly or re-deploy from scratch.\n"} 2021-03-01 21:35:36,732+0600 ERROR otopi.context context._executeMethod:154 Failed to execute stage 'Closing up': Failed executing ansible-playbook 2021-03-01 21:35:46,159+0600 ERROR otopi.ovirt_hosted_engine_setup.ansible_utils ansible_utils._process_output:109 fatal: [localhost]: UNREACHABLE! => {"changed": false, "msg": "Failed to connect to the host via ssh: ssh: connect to host ovirt4-adm.domain.local port 22: No route to host", "skip_reason": "Host localhost is unreachable", "unreachable": true} 2021-03-01 21:36:08,829+0600 ERROR otopi.plugins.gr_he_common.core.misc misc._terminate:167 Hosted Engine deployment failed: please check the logs for the issue, fix accordingly or re-deploy from scratch. [root@node02 ~]$ ip a s dev virbr0 27: virbr0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state UP group default qlen 1000 link/ether 52:54:00:e8:50:9e brd ff:ff:ff:ff:ff:ff inet 192.168.1.1/24 brd 192.168.1.255 scope global virbr0 valid_lft forever preferred_lft forever [root@node02 ~]$ ip neigh s dev virbr0 192.168.1.207 lladdr 00:16:3e:7c:77:ce REACHABLE Naturally, Ovirt Engine is not available via ssh. his ip should be 10.11.120.144 ovirt4-adm.domain.local, but he is [root @ ovirt4-adm ~] # cat / etc / hosts 127.0.0.1 localhost localhost.localdomain localhost4 localhost4.localdomain4 :: 1 localhost localhost.localdomain localhost6 localhost6.localdomain6 10.11.120.142 node02.domain.local 192.168.1.207 ovirt4-adm.domain.local # hosted-engine-setup- / var / tmp / localvmsualu32r where to dig, watch. tell me what to do next.

The installation of the ovirt "hosted-engine" hangs at the stage "[INFO] TASK [ovirt.ovirt.hosted_engine_setup: Wait for the host to be up]" (https://pastebin.com/zvf9T8nP) for 20 minutes, then gives an error " [ERROR] fatal: [localhost]: FAILED! => {"Changed": false, "msg": Host is not up, please check logs, perhaps also on the engine machine "}". At the same time vm with "engine" started, but the installation was not complete. In the logs on "ovirt-engine vm" https://pastebin.com/95GCVeQp I see that the ansible script searches the host for node02 python2: "2021-03-06 19:42:30 OMST - fatal: [node02.locdomain.local]: FAILED! => {" changed ": false," module_stderr ":" / bin / sh: / usr / bin / python2: No such file or directory \ n "," module_stdout ":" "," msg ":" The module failed to execute correctly, you probably need to set the interpreter. \ nSee stdout / stderr for the exact error "," rc ": 127}" the choice of the version is carried out according to the condition: - name: Detect host operating system set_fact: el_ver: "{{ ansible_distribution_major_version|int if ansible_distribution == 'RedHat' or ansible_distribution == 'CentOS' else 0 }}" fc_ver: "{{ ansible_distribution_major_version|int if ansible_distribution == 'Fedora' else 0 }}" "remote_addr" : "node02.locdomain.local", "res" : { "changed" : false, "ansible_facts" : { "el_ver" : "0", "fc_ver" : "0" }, ansible_python_interpreter : "{{ '/usr/bin/python3' if (fc_ver|int > 0 or el_ver|int >= 8) else '/usr/bin/python2' }} /etc/os-release NAME="AlmaLinux" VERSION="8.3 (Purple Manul)" ID="almalinux" ID_LIKE="rhel centos fedora" VERSION_ID="8.3" PLATFORM_ID="platform:el8" PRETTY_NAME="AlmaLinux 8.3 RC (Purple Manul)" ANSI_COLOR="0;34" CPE_NAME="cpe:/o:almalinux:almalinux:8.3:rc" HOME_URL="https://almalinux.org/" BUG_REPORT_URL="https://bugs.almalinux.org/" cat /etc/redhat-release AlmaLinux 8.3 RC (Purple Manul)

On Tue, Mar 9, 2021 at 7:33 AM <grig.4n@gmail.com> wrote:
The installation of the ovirt "hosted-engine" hangs at the stage "[INFO] TASK [ovirt.ovirt.hosted_engine_setup: Wait for the host to be up]" (https://pastebin.com/zvf9T8nP) for 20 minutes, then gives an error " [ERROR] fatal: [localhost]: FAILED! => {"Changed": false, "msg": Host is not up, please check logs, perhaps also on the engine machine "}". At the same time vm with "engine" started, but the installation was not complete. In the logs on "ovirt-engine vm" https://pastebin.com/95GCVeQp I see that the ansible script searches the host for node02 python2: "2021-03-06 19:42:30 OMST - fatal: [node02.locdomain.local]: FAILED! => {" changed ": false," module_stderr ":" / bin / sh: / usr / bin / python2: No such file or directory \ n "," module_stdout ":" "," msg ":" The module failed to execute correctly, you probably need to set the interpreter. \ nSee stdout / stderr for the exact error "," rc ": 127}" the choice of the version is carried out according to the condition:
- name: Detect host operating system set_fact: el_ver: "{{ ansible_distribution_major_version|int if ansible_distribution == 'RedHat' or ansible_distribution == 'CentOS' else 0 }}" fc_ver: "{{ ansible_distribution_major_version|int if ansible_distribution == 'Fedora' else 0 }}"
"remote_addr" : "node02.locdomain.local", "res" : { "changed" : false, "ansible_facts" : { "el_ver" : "0", "fc_ver" : "0" }, ansible_python_interpreter : "{{ '/usr/bin/python3' if (fc_ver|int > 0 or el_ver|int >= 8) else '/usr/bin/python2' }}
/etc/os-release NAME="AlmaLinux" VERSION="8.3 (Purple Manul)" ID="almalinux" ID_LIKE="rhel centos fedora" VERSION_ID="8.3" PLATFORM_ID="platform:el8" PRETTY_NAME="AlmaLinux 8.3 RC (Purple Manul)" ANSI_COLOR="0;34" CPE_NAME="cpe:/o:almalinux:almalinux:8.3:rc" HOME_URL="https://almalinux.org/" BUG_REPORT_URL="https://bugs.almalinux.org/"
cat /etc/redhat-release AlmaLinux 8.3 RC (Purple Manul)
Nice catch! Can you please check (I think you can see this in the logs) what ansible says ansible_distribution and ansible_distribution_major_version are, for you? Best regards, -- Didi

On Tue, Mar 9, 2021 at 8:17 AM Yedidyah Bar David <didi@redhat.com> wrote:
On Tue, Mar 9, 2021 at 7:33 AM <grig.4n@gmail.com> wrote:
The installation of the ovirt "hosted-engine" hangs at the stage "[INFO] TASK [ovirt.ovirt.hosted_engine_setup: Wait for the host to be up]" (https://pastebin.com/zvf9T8nP) for 20 minutes, then gives an error " [ERROR] fatal: [localhost]: FAILED! => {"Changed": false, "msg": Host is not up, please check logs, perhaps also on the engine machine "}". At the same time vm with "engine" started, but the installation was not complete. In the logs on "ovirt-engine vm" https://pastebin.com/95GCVeQp I see that the ansible script searches the host for node02 python2: "2021-03-06 19:42:30 OMST - fatal: [node02.locdomain.local]: FAILED! => {" changed ": false," module_stderr ":" / bin / sh: / usr / bin / python2: No such file or directory \ n "," module_stdout ":" "," msg ":" The module failed to execute correctly, you probably need to set the interpreter. \ nSee stdout / stderr for the exact error "," rc ": 127}" the choice of the version is carried out according to the condition:
- name: Detect host operating system set_fact: el_ver: "{{ ansible_distribution_major_version|int if ansible_distribution == 'RedHat' or ansible_distribution == 'CentOS' else 0 }}" fc_ver: "{{ ansible_distribution_major_version|int if ansible_distribution == 'Fedora' else 0 }}"
"remote_addr" : "node02.locdomain.local", "res" : { "changed" : false, "ansible_facts" : { "el_ver" : "0", "fc_ver" : "0" }, ansible_python_interpreter : "{{ '/usr/bin/python3' if (fc_ver|int > 0 or el_ver|int >= 8) else '/usr/bin/python2' }}
/etc/os-release NAME="AlmaLinux" VERSION="8.3 (Purple Manul)" ID="almalinux" ID_LIKE="rhel centos fedora" VERSION_ID="8.3" PLATFORM_ID="platform:el8" PRETTY_NAME="AlmaLinux 8.3 RC (Purple Manul)" ANSI_COLOR="0;34" CPE_NAME="cpe:/o:almalinux:almalinux:8.3:rc" HOME_URL="https://almalinux.org/" BUG_REPORT_URL="https://bugs.almalinux.org/"
cat /etc/redhat-release AlmaLinux 8.3 RC (Purple Manul)
Nice catch!
Can you please check (I think you can see this in the logs) what ansible says ansible_distribution and ansible_distribution_major_version are, for you?
BTW, to clarify: The answer to the above question should allow a very quick and simple patch that will fix the _above_ issue. Assuming that AlmaLinux is compatible enough with RHEL, providing full support includes at least: 1. Finding all such places and fix them... A first approximation is searching the sources for 'centos' (case insensitive), linux_distribution (for python code), probably a few others. 2. Answering some harder questions, such as what the intention is, regarding ovirt-appliance and ovirt-node. At some point, I suppose we'll build them based on CentOS Stream. People that want AlmaLinux (just as an example - applies similarly to any other RHEL rebuild) for everything, and also want to have/use appliance/node, probably also want to build them based on AlmaLinux. This is not just providing some patch for issues like the above one. 3. Also consider non-OS-provided 3rd-party dependencies. Including e.g. wildfly, gluster, etc. For some of them, the existing builds that oVirt uses for CentOS might be good enough as-is, but not sure if this covers everything. Best regards, -- Didi

"ansible_distribution" : "AlmaLinux", "ansible_distribution_release" : "Purple Manul", "ansible_distribution_version" : "8.3", "ansible_distribution_major_version" : "8", "ansible_distribution_file_path" : "/etc/redhat-release", "ansible_distribution_file_variety" : "RedHat", "ansible_distribution_file_parsed" : true, "ansible_os_family" : "AlmaLinux"

On Tue, Mar 9, 2021 at 8:44 AM <grig.4n@gmail.com> wrote:
"ansible_distribution" : "AlmaLinux", "ansible_distribution_release" : "Purple Manul", "ansible_distribution_version" : "8.3", "ansible_distribution_major_version" : "8", "ansible_distribution_file_path" : "/etc/redhat-release", "ansible_distribution_file_variety" : "RedHat", "ansible_distribution_file_parsed" : true, "ansible_os_family" : "AlmaLinux"
Great. Would you like to try to fix this? The code is in the 'ovirt-engine' git repo, see also [1]. If not, perhaps file a bug. Thanks. [1] https://www.ovirt.org/develop/dev-process/working-with-gerrit.html Best regards, -- Didi
participants (3)
-
Grig C
-
grig.4n@gmail.com
-
Yedidyah Bar David