On Wed, Jan 31, 2018 at 11:48 AM, Simone Tiraboschi <stirabos@redhat.com> wrote:


Ciao Gianluca,
we have an issue logging messages with special unicode chars from ansible, it's tracked here:
but this is just hiding your real issue.

I'm almost sure that you are facing an issue writing on NFS and thwn dd returns us an error message with \u2018 and \u2019.
Can you please check your NFS permissions?
 

Ciao Simone, thanks for answering.
I think you were right.
Previously I had this:

/nfs/SHE_DOMAIN *(rw)

Now I have changed to:

/nfs/SHE_DOMAIN *(rw,anonuid=36,anongid=36,all_squash)

I restarted the deploy with the answer file

# hosted-engine --deploy --config-append=/var/lib/ovirt-hosted-engine-setup/answers/answers-20180129164431.conf

and it went ahead... and I have contents inside the directory:

# ll /nfs/SHE_DOMAIN/a0351a82-734d-4d9a-a75e-3313d2ffe23a/
total 12
drwxr-xr-x. 2 vdsm kvm 4096 Jan 29 16:40 dom_md
drwxr-xr-x. 6 vdsm kvm 4096 Jan 29 16:43 images
drwxr-xr-x. 4 vdsm kvm 4096 Jan 29 16:40 master

But it ended with a problem regarding engine vm:

[ INFO  ] TASK [Wait for engine to start]
[ INFO  ] ok: [localhost]
[ INFO  ] TASK [Set engine pub key as authorized key without validating the TLS/SSL certificates]
[ INFO  ] changed: [localhost]
[ INFO  ] TASK [Force host-deploy in offline mode]
[ INFO  ] changed: [localhost]
[ INFO  ] TASK [include_tasks]
[ INFO  ] ok: [localhost]
[ INFO  ] TASK [Obtain SSO token using username/password credentials]
[ INFO  ] ok: [localhost]
[ INFO  ] TASK [Add host]
[ INFO  ] changed: [localhost]
[ INFO  ] TASK [Wait for the host to become non operational]
[ INFO  ] ok: [localhost]
[ INFO  ] TASK [Get virbr0 routing configuration]
[ INFO  ] changed: [localhost]
[ INFO  ] TASK [Get ovirtmgmt route table id]
[ INFO  ] changed: [localhost]
[ INFO  ] TASK [Check network configuration]
[ INFO  ] changed: [localhost]
[ INFO  ] TASK [Clean network configuration]
[ INFO  ] changed: [localhost]
[ INFO  ] TASK [Restore network configuration]
[ INFO  ] changed: [localhost]
[ INFO  ] TASK [Wait for the host to be up]
[ ERROR ] Error: Failed to read response.
[ ERROR ] fatal: [localhost]: FAILED! => {"attempts": 50, "changed": false, "msg": "Failed to read response."}
[ ERROR ] Failed to execute stage 'Closing up': Failed executing ansible-playbook
[ INFO  ] Stage: Clean up
[ INFO  ] Cleaning temporary resources
[ INFO  ] TASK [Gathering Facts]
[ INFO  ] ok: [localhost]
[ INFO  ] TASK [Remove local vm dir]
[ INFO  ] ok: [localhost]
[ INFO  ] Generating answer file '/var/lib/ovirt-hosted-engine-setup/answers/answers-20180201104600.conf'
[ INFO  ] Stage: Pre-termination
[ INFO  ] Stage: Termination
[ ERROR ] Hosted Engine deployment failed: this system is not reliable, please check the issue,fix and redeploy
          Log file is located at /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20180201102603-1of5a1.log

Under /var/log/libvirt/qemu of host from where I'm running the hosted-engine deploy I see this


2018-02-01 09:29:05.515+0000: starting up libvirt version: 3.2.0, package: 14.el7_4.7 (CentOS BuildSystem <http://bugs.centos.org>, 2018-01-04-19:31:34, c1bm.rdu2.centos.org), qemu version: 2.9.0(qemu-kvm-ev-2.9.0-16.el7_4.13.1), hostname: ov42.mydomain
LC_ALL=C PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin QEMU_AUDIO_DRV=none /usr/libexec/qemu-kvm -name guest=HostedEngineLocal,debug-threads=on -S -object secret,id=masterKey0,format=raw,file=/var/lib/libvirt/qemu/domain-1-HostedEngineLocal/master-key.aes -machine pc-i440fx-rhel7.4.0,accel=kvm,usb=off,dump-guest-core=off -cpu Westmere,+kvmclock -m 6184 -realtime mlock=off -smp 1,sockets=1,cores=1,threads=1 -uuid 8c8f8163-5b69-4ff5-b67c-07b1a9b8f100 -no-user-config -nodefaults -chardev socket,id=charmonitor,path=/var/lib/libvirt/qemu/domain-1-HostedEngineLocal/monitor.sock,server,nowait -mon chardev=charmonitor,id=monitor,mode=control -rtc base=utc -no-shutdown -global PIIX4_PM.disable_s3=1 -global PIIX4_PM.disable_s4=1 -boot menu=off,strict=on -device virtio-serial-pci,id=virtio-serial0,bus=pci.0,addr=0x4 -drive file=/var/tmp/localvm1ClXud/images/918bbfc1-d599-4170-9a92-1ac417bf7658/bb8b3078-fddb-4ce3-8da0-0a191768a357,format=qcow2,if=none,id=drive-virtio-disk0 -device virtio-blk-pci,scsi=off,bus=pci.0,addr=0x5,drive=drive-virtio-disk0,id=virtio-disk0,bootindex=1 -drive file=/var/tmp/localvm1ClXud/seed.iso,format=raw,if=none,id=drive-ide0-0-0,readonly=on -device ide-cd,bus=ide.0,unit=0,drive=drive-ide0-0-0,id=ide0-0-0 -netdev tap,fd=26,id=hostnet0,vhost=on,vhostfd=28 -device virtio-net-pci,netdev=hostnet0,id=net0,mac=00:16:3e:15:7b:27,bus=pci.0,addr=0x3 -chardev pty,id=charserial0 -device isa-serial,chardev=charserial0,id=serial0 -chardev socket,id=charchannel0,path=/var/lib/libvirt/qemu/channel/target/domain-1-HostedEngineLocal/org.qemu.guest_agent.0,server,nowait -device virtserialport,bus=virtio-serial0.0,nr=1,chardev=charchannel0,id=channel0,name=org.qemu.guest_agent.0 -vnc 127.0.0.1:0 -device VGA,id=video0,vgamem_mb=16,bus=pci.0,addr=0x2 -object rng-random,id=objrng0,filename=/dev/random -device virtio-rng-pci,rng=objrng0,id=rng0,bus=pci.0,addr=0x6 -msg timestamp=on
2018-02-01T09:29:05.771459Z qemu-kvm: -chardev pty,id=charserial0: char device redirected to /dev/pts/3 (label charserial0)
2018-02-01T09:34:19.445774Z qemu-kvm: terminating on signal 15 from pid 6052 (/usr/sbin/libvirtd)
2018-02-01 09:34:19.668+0000: shutting down, reason=shutdown

In /var/log/messages:

Feb  1 10:29:05 ov42 systemd-machined: New machine qemu-1-HostedEngineLocal.
Feb  1 10:29:05 ov42 systemd: Started Virtual Machine qemu-1-HostedEngineLocal.
Feb  1 10:29:05 ov42 systemd: Starting Virtual Machine qemu-1-HostedEngineLocal.
Feb  1 10:29:05 ov42 kvm: 1 guest now active
Feb  1 10:29:06 ov42 python: ansible-command Invoked with warn=True executable=None _uses_shell=True
 _raw_params=virsh -r net-dhcp-leases default | grep -i 00:16:3e:15:7b:27 | awk '{ print $5 }' | cut
 -f1 -d'/' removes=None creates=None chdir=None stdin=None
Feb  1 10:29:07 ov42 kernel: virbr0: port 2(vnet0) entered learning state
Feb  1 10:29:09 ov42 kernel: virbr0: port 2(vnet0) entered forwarding state
Feb  1 10:29:09 ov42 kernel: virbr0: topology change detected, propagating
Feb  1 10:29:09 ov42 NetworkManager[749]: <info>  [1517477349.5180] device (virbr0): link connected
Feb  1 10:29:16 ov42 python: ansible-command Invoked with warn=True executable=None _uses_shell=True
 _raw_params=virsh -r net-dhcp-leases default | grep -i 00:16:3e:15:7b:27 | awk '{ print $5 }' | cut
 -f1 -d'/' removes=None creates=None chdir=None stdin=None
Feb  1 10:29:27 ov42 python: ansible-command Invoked with warn=True executable=None _uses_shell=True
 _raw_params=virsh -r net-dhcp-leases default | grep -i 00:16:3e:15:7b:27 | awk '{ print $5 }' | cut
 -f1 -d'/' removes=None creates=None chdir=None stdin=None
Feb  1 10:29:30 ov42 dnsmasq-dhcp[6322]: DHCPDISCOVER(virbr0) 00:16:3e:15:7b:27
Feb  1 10:29:30 ov42 dnsmasq-dhcp[6322]: DHCPOFFER(virbr0) 192.168.122.200 00:16:3e:15:7b:27
Feb  1 10:29:30 ov42 dnsmasq-dhcp[6322]: DHCPREQUEST(virbr0) 192.168.122.200 00:16:3e:15:7b:27
Feb  1 10:29:30 ov42 dnsmasq-dhcp[6322]: DHCPACK(virbr0) 192.168.122.200 00:16:3e:15:7b:27
. . .
Feb  1 10:34:00 ov42 systemd: Starting Virtualization daemon...
Feb  1 10:34:00 ov42 python: ansible-ovirt_hosts_facts Invoked with pattern=name=ov42.mydomain status=up fetch_nested=False nested_attributes=[] auth={'ca_file': None, 'url': 'https://ov42she.mydomain/ovirt-engine/api', 'insecure': True, 'kerberos': False, 'compress': True, 'headers': None, 'token': 'GOK2wLFZ0PIs1GbXVQjNW-yBlUtZoGRa2I92NkCkm6lwdlQV-dUdP5EjInyGGN_zEVEHFKgR6nuZ-eIlfaM_lw', 'timeout': 0}
Feb  1 10:34:03 ov42 systemd: Started Virtualization daemon.
Feb  1 10:34:03 ov42 systemd: Reloading.
Feb  1 10:34:03 ov42 systemd: [/usr/lib/systemd/system/ip6tables.service:3] Failed to add dependency on syslog.target,iptables.service, ignoring: Invalid argument
Feb  1 10:34:03 ov42 systemd: Cannot add dependency job for unit lvm2-lvmetad.socket, ignoring: Unit is masked.
Feb  1 10:34:03 ov42 systemd: Starting Cockpit Web Service...
Feb  1 10:34:03 ov42 dnsmasq[6322]: read /etc/hosts - 4 addresses
Feb  1 10:34:03 ov42 dnsmasq[6322]: read /var/lib/libvirt/dnsmasq/default.addnhosts - 0 addresses
Feb  1 10:34:03 ov42 dnsmasq-dhcp[6322]: read /var/lib/libvirt/dnsmasq/default.hostsfile
Feb  1 10:34:03 ov42 systemd: Started Cockpit Web Service.
Feb  1 10:34:03 ov42 cockpit-ws: Using certificate: /etc/cockpit/ws-certs.d/0-self-signed.cert
Feb  1 10:34:03 ov42 libvirtd: 2018-02-01 09:34:03.840+0000: 6076: info : libvirt version: 3.2.0, package: 14.el7_4.7 (CentOS BuildSystem <http://bugs.centos.org>, 2018-01-04-19:31:34, c1bm.rdu2.centos.org)
Feb  1 10:34:03 ov42 libvirtd: 2018-02-01 09:34:03.840+0000: 6076: info : hostname: ov42.mydomain
Feb  1 10:34:03 ov42 libvirtd: 2018-02-01 09:34:03.840+0000: 6076: error : virDirOpenInternal:2829 : cannot open directory '/var/tmp/localvm7I0SSJ/images/918bbfc1-d599-4170-9a92-1ac417bf7658': No such file or directory
Feb  1 10:34:03 ov42 libvirtd: 2018-02-01 09:34:03.841+0000: 6076: error : storageDriverAutostart:204 : internal error: Failed to autostart storage pool '918bbfc1-d599-4170-9a92-1ac417bf7658': cannot open directory '/var/tmp/localvm7I0SSJ/images/918bbfc1-d599-4170-9a92-1ac417bf7658': No such file or directory
Feb  1 10:34:03 ov42 libvirtd: 2018-02-01 09:34:03.841+0000: 6076: error : virDirOpenInternal:2829 : cannot open directory '/var/tmp/localvm7I0SSJ': No such file or directory
Feb  1 10:34:03 ov42 libvirtd: 2018-02-01 09:34:03.841+0000: 6076: error : storageDriverAutostart:204 : internal error: Failed to autostart storage pool 'localvm7I0SSJ': cannot open directory '/var/tmp/localvm7I0SSJ': No such file or directory
Feb  1 10:34:03 ov42 systemd: Stopping Suspend/Resume Running libvirt Guests...
Feb  1 10:34:04 ov42 libvirt-guests.sh: Running guests on qemu+tls://ov42.mydomain/system URI: HostedEngineLocal
Feb  1 10:34:04 ov42 libvirt-guests.sh: Shutting down guests on qemu+tls://ov42.mydomain/system URI...
Feb  1 10:34:04 ov42 libvirt-guests.sh: Starting shutdown on guest: HostedEngineLocal

If I understood corrctly it seems that libvirtd took in charge the ip assignement, using the default 192.168.122.x network, while my host and my engine should be on 10.4.4.x...??
Currently on host, after the failed deploy, I have:

# brctl show
bridge name bridge id STP enabled interfaces
;vdsmdummy; 8000.000000000000 no
ovirtmgmt 8000.001a4a17015d no eth0
virbr0 8000.52540084b832 yes virbr0-nic

BTW: on host I have network managed by NetworkManager. It is supported now in upcoming 4.2.1, isn't it?

Gianluca