[snip]
Snippet for hosted engine hosts shutdown:
- name: Shutdown of HE hosts
command: >-
ssh -o UserKnownHostsFile=/dev/null -o StrictHostKeyChecking=no
-i /etc/pki/ovirt-engine/keys/engine_id_rsa -p {{ item.ssh.port }}
-t root@{{ item.address }} '{{ he_shutdown_cmd }}'
async: 1000
poll: 0
with_items:
- "{{ he_hosts }}"
where the he_shutdown_cmd var is defined as:
he_shutdown_cmd: >-
while hosted-engine --vm-status | grep "\"vm\": \"up\"" >/dev/null;
do sleep 5;
done;
sanlock client shutdown -f 1;
shutdown -h now
Snippet for the Engine VM shutdown
- name: Shutdown engine host/VM
command: shutdown -h now
async: 1000
poll: 0
[snip]
Could it be useful to insert the /usr/share/glusterfs/scripts/stop-all-gluster-processes.sh command, as suggested by Strahil, after the sanlock one, in case of GlusterFS domain?
Also one further note:
inside the role it is used the ovirt_host_facts module. I get this when I use it and debug its content:
{
"msg": "The 'ovirt_host_facts' module has been renamed to 'ovirt_host_info', and the renamed one no longer returns ansible_facts",
"version": "2.13"
}
So perhaps it should be considered to change and use ovirt_host_info instead? Any plan?
Thanks for reading.
Gianluca
Hello,
I would like to keep on with this to have a better experience.
Environment is physical 4.3.10 single host HCI that shows the same problems as above.
So I modified the role file adding after sanlock shutdown the gluster stop script
[root@ovengine tasks]# pwd
/root/roles/ovirt.shutdown_env/tasks
[root@ovengine tasks]# diff main.yml main.yml.orig
79d78
< /usr/share/glusterfs/scripts/stop-all-gluster-processes.sh
[root@ovengine tasks]#
Now the poweroff completes, even if get these errors about stopping swap and gluster bricks filesystems:
When I power on again the server in a second time, the environment starts ok in global maintenance and when I exit it, engine and gluster volumes start ok.
Even if there is a 2-3 minutes delay (I already opened a thread about this) from the moment when you see the storage domains up in the web admin gui and the moment when they are truely up (file systems /rhev/data-center/mnt/glusterSD/... mounted). So if in the mean time you try to start a VM you get an error because disks not found...
Comments?
Gianluca