
On Mon, May 25, 2020 at 8:50 AM Yedidyah Bar David <didi@redhat.com <mailto:didi@redhat.com>> wrote:
On Sun, May 24, 2020 at 9:36 PM Gianluca Cecchi <gianluca.cecchi@gmail.com <mailto:gianluca.cecchi@gmail.com>> wrote: > > On Sun, May 24, 2020 at 11:47 AM Yedidyah Bar David <didi@redhat.com <mailto:didi@redhat.com>> wrote: >> >> >> >> Hi, Gianluca. Replying to your email on "4.4 HCI Install Failure - >> Missing /etc/pki/CA/cacert.pem": >> >> On Sun, May 24, 2020 at 12:28 PM Gianluca Cecchi >> <gianluca.cecchi@gmail.com <mailto:gianluca.cecchi@gmail.com>> wrote: >> > >> > I I remember correctly it happened to me during the beta cycle and the only "strange" character I used for the admin password was the @ >> > Donna if it related with what you reported for the % character >> >> Did you open a bug? >> >> In any case, my above patch is not supposed to fix '@', only '%' (I think). >> >> Thanks and best regards, >> > > No, I didn't open a bug, because I scratched the system and installed again this time without the error, but I don't remember if I used the same password with the @ character or not.... > I will put attention in case of future 4.4 new installations
Very well, thanks :-) -- Didi
Just to avoid opening a bug for a different thing, today I tried a single host HCI setup with the wizard and it failed. Installed from ovirt-node-ng final 4.4 iso. I see I have no /etc/pki/CA directory on the host at the moment, but I don't know if the install workflow had not arrived there yet or what. Last lines in my wizard window are these ones below. I have had this too and saw that during the test week is was also reported. The 'solution' was to run /usr/sbin/ovirt-hosted-engine-cleanup and let
On 3-6-2020 15:29, Gianluca Cecchi wrote: the wizard stay where it is and afterwards click Redeploy VM and then it will run to its conclusion. Don't know what precisely it fixes. Had a look and couldn't find it. Joop
Password used contains only letters, numbers and the "_" character in this attempt I'm in the "Prepare VM" stage.
[ INFO ] TASK [ovirt.hosted_engine_setup : Stop libvirt service] [ INFO ] changed: [localhost] [ INFO ] TASK [ovirt.hosted_engine_setup : Drop vdsm config statements] [ INFO ] changed: [localhost] [ INFO ] TASK [ovirt.hosted_engine_setup : Restore initial abrt config files] [ INFO ] changed: [localhost] [ INFO ] TASK [ovirt.hosted_engine_setup : Restart abrtd service] [ INFO ] changed: [localhost] [ INFO ] TASK [ovirt.hosted_engine_setup : Drop libvirt sasl2 configuration by vdsm] [ INFO ] changed: [localhost] [ INFO ] TASK [ovirt.hosted_engine_setup : Stop and disable services] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.hosted_engine_setup : Restore initial libvirt default network configuration] [ INFO ] changed: [localhost] [ INFO ] TASK [ovirt.hosted_engine_setup : Start libvirt] [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": "Unable to start service libvirtd: Job for libvirtd.service failed because the control process exited with error code.\nSee \"systemctl status libvirtd.service\" and \"journalctl -xe\" for details.\n"}
Status of libvirtd service is this one:
[root@ovirt01 g.cecchi]# systemctl status libvirtd -l --no-pager ● libvirtd.service - Virtualization daemon Loaded: loaded (/usr/lib/systemd/system/libvirtd.service; enabled; vendor preset: enabled) Drop-In: /etc/systemd/system/libvirtd.service.d └─unlimited-core.conf Active: failed (Result: exit-code) since Wed 2020-06-03 15:13:35 CEST; 7min ago Docs: man:libvirtd(8) https://libvirt.org Process: 20001 ExecStart=/usr/sbin/libvirtd $LIBVIRTD_ARGS (code=exited, status=6) Main PID: 20001 (code=exited, status=6) Tasks: 2 (limit: 32768) Memory: 70.1M CGroup: /system.slice/libvirtd.service ├─3926 /usr/sbin/dnsmasq --conf-file=/var/lib/libvirt/dnsmasq/default.conf --leasefile-ro --dhcp-script=/usr/libexec/libvirt_leaseshelper └─3927 /usr/sbin/dnsmasq --conf-file=/var/lib/libvirt/dnsmasq/default.conf --leasefile-ro --dhcp-script=/usr/libexec/libvirt_leaseshelper
Jun 03 15:13:35 ovirt01.mydomain.local systemd[1]: libvirtd.service: Service RestartSec=100ms expired, scheduling restart. Jun 03 15:13:35 ovirt01.mydomain.local systemd[1]: libvirtd.service: Scheduled restart job, restart counter is at 5. Jun 03 15:13:35 ovirt01.mydomain.local systemd[1]: Stopped Virtualization daemon. Jun 03 15:13:35 ovirt01.mydomain.local systemd[1]: libvirtd.service: Start request repeated too quickly. Jun 03 15:13:35 ovirt01.mydomain.local systemd[1]: libvirtd.service: Failed with result 'exit-code'. Jun 03 15:13:35 ovirt01.mydomain.local systemd[1]: Failed to start Virtualization daemon. [root@ovirt01 g.cecchi]#
Let me know what files do you want to analyze the problem Under /var/log/ovirt-hosted-engine-setup I have:
[root@ovirt01 ovirt-hosted-engine-setup]# ls -lrt total 632 -rw-r--r--. 1 root root 123814 Jun 3 15:07 ovirt-hosted-engine-setup-ansible-get_network_interfaces-20205315737-ooohyb.log -rw-r--r--. 1 root root 126674 Jun 3 15:08 ovirt-hosted-engine-setup-ansible-validate_hostnames-20205315737-oqixuw.log -rw-r--r--. 1 root root 127548 Jun 3 15:10 ovirt-hosted-engine-setup-ansible-validate_hostnames-202053151022-yls4qo.log -rw-r--r--. 1 root root 261482 Jun 3 15:13 ovirt-hosted-engine-setup-ansible-initial_clean-20205315123-7x25zv.log [root@ovirt01 ovirt-hosted-engine-setup]#
In this environment no dns but entry in /etc/hosts of the server. Host is on 192.168.1.x on eno1 and a vlan 100 on the same interface, used for "simulated" storage network
2: eno1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc fq_codel state UP group default qlen 1000 link/ether b8:ae:ed:7f:17:11 brd ff:ff:ff:ff:ff:ff inet 192.168.1.211/24 <http://192.168.1.211/24> brd 192.168.1.255 scope global noprefixroute eno1 valid_lft forever preferred_lft forever inet6 fe80::315c:2e9f:d009:8c0c/64 scope link noprefixroute valid_lft forever preferred_lft forever
7: eno1.100@eno1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state UP group default qlen 1000 link/ether b8:ae:ed:7f:17:11 brd ff:ff:ff:ff:ff:ff inet 192.168.100.211/24 <http://192.168.100.211/24> brd 192.168.100.255 scope global noprefixroute eno1.100 valid_lft forever preferred_lft forever inet6 fe80::be0f:c4be:cf7a:9da5/64 scope link noprefixroute valid_lft forever preferred_lft forever
Thanks, Gianluca
_______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-leave@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/IPDND2OXUV2DXL...