On Wed, Apr 15, 2020 at 4:52 PM Gianluca Cecchi <gianluca.cecchi(a)gmail.com>
wrote:
[snip]
Snippet for hosted engine hosts shutdown:
- name: Shutdown of HE hosts
command: >-
ssh -o UserKnownHostsFile=/dev/null -o StrictHostKeyChecking=no
-i /etc/pki/ovirt-engine/keys/engine_id_rsa -p {{
item.ssh.port }}
-t root@{{ item.address }} '{{ he_shutdown_cmd }}'
async: 1000
poll: 0
with_items:
- "{{ he_hosts }}"
where the he_shutdown_cmd var is defined as:
he_shutdown_cmd: >-
while hosted-engine --vm-status | grep "\"vm\":
\"up\""
>/dev/null;
do sleep 5;
done;
sanlock client shutdown -f 1;
shutdown -h now
Snippet for the Engine VM shutdown
- name: Shutdown engine host/VM
command: shutdown -h now
async: 1000
poll: 0
[snip]
Could it be useful to insert the
/usr/share/glusterfs/scripts/stop-all-gluster-processes.sh command, as
suggested by Strahil, after the sanlock one, in case of GlusterFS domain?
Also one further note:
inside the role it is used the ovirt_host_facts module. I get this when I
use it and debug its content:
{
"msg": "The 'ovirt_host_facts' module has been
renamed to
'ovirt_host_info', and the renamed one no longer returns ansible_facts",
"version": "2.13"
}
So perhaps it should be considered to change and use ovirt_host_info
instead? Any plan?
Thanks for reading.
Gianluca
Hello,
I would like to keep on with this to have a better experience.
Environment is physical 4.3.10 single host HCI that shows the same problems
as above.
So I modified the role file adding after sanlock shutdown the gluster stop
script
[root@ovengine tasks]# pwd
/root/roles/ovirt.shutdown_env/tasks
[root@ovengine tasks]# diff main.yml main.yml.orig
79d78
< /usr/share/glusterfs/scripts/stop-all-gluster-processes.sh
[root@ovengine tasks]#
Now the poweroff completes, even if get these errors about stopping swap
and gluster bricks filesystems:
https://drive.google.com/file/d/1oh0sNC3ta5qP0KAcibTdDc5N_lpil8pS/view?us...
When I power on again the server in a second time, the environment starts
ok in global maintenance and when I exit it, engine and gluster volumes
start ok.
Even if there is a 2-3 minutes delay (I already opened a thread about this)
from the moment when you see the storage domains up in the web admin gui
and the moment when they are truely up (file systems
/rhev/data-center/mnt/glusterSD/... mounted). So if in the mean time you
try to start a VM you get an error because disks not found...
Comments?
Gianluca