Now that I've been able to get past my issues with q35 bios I decided to try the
hyperconverged install. setting up the storage was no problem and the engine setup went
fine until the point where it wanted to use that glusterfs storage. I got the error,
pretty quickly I might add, of
2020-06-30 02:17:50,689-0400 DEBUG ansible on_any args
<ansible.executor.task_result.TaskResult object at 0x7f83f85b4278> kwargs
2020-06-30 02:17:51,428-0400 INFO ansible task start {'status': 'OK',
'ansible_type': 'task', 'ansible_playbook':
'/usr/share/ovirt-hosted-engine-setup/ansible/trigger_role.yml',
'ansible_task': 'ovirt.hosted_engine_setup : Add glusterfs storage
domain'}
2020-06-30 02:17:51,428-0400 DEBUG ansible on_any args TASK: ovirt.hosted_engine_setup :
Add glusterfs storage domain kwargs is_conditional:False
2020-06-30 02:17:51,429-0400 DEBUG ansible on_any args localhostTASK:
ovirt.hosted_engine_setup : Add glusterfs storage domain kwargs
2020-06-30 02:17:53,656-0400 DEBUG var changed: host "localhost" var
"otopi_storage_domain_details_gluster" type "<class
'dict'>" value: "{
"changed": false,
"exception": "Traceback (most recent call last):
File
\"/tmp/ansible_ovirt_storage_domain_payload_jh42k_ip/ansible_ovirt_storage_domain_payload.zip/ansible/modules/cloud/ovirt/ovirt_storage_domain.py\",
line 792, in main
File
\"/tmp/ansible_ovirt_storage_domain_payload_jh42k_ip/ansible_ovirt_storage_domain_payload.zip/ansible/module_utils/ovirt.py\",
line 623, in create
**kwargs
File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/services.py\", line 26097,
in add
return self._internal_add(storage_domain, headers, query, wait)
File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 232, in
_internal_add
return future.wait() if wait else future
File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 55, in
wait
return self._code(response)
File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 229, in
callback
self._check_fault(response)
File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 132, in
_check_fault
self._raise_error(response, body)
File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 118, in
_raise_error
raise error\novirtsdk4.Error: Fault reason is \"Operation Failed\". Fault
detail is \"[Failed to fetch Gluster Volume List]\". HTTP response code is
400.\n",
"failed": true,
"msg": "Fault reason is \"Operation Failed\". Fault detail is
\"[Failed to fetch Gluster Volume List]\". HTTP response code is 400."
}"
The glusterfs cli worked fine by that point, had no issues with volume list, so I went
into the node:6900/ovirt-engine forwarder to the appliance and could see that at the time
of that error in the event logs there was a message that it was starting to update things
and a few seconds later a message that the volumes were available.
I am thinking that there is need for a retry loop in this step or that something in the
path isn't waiting for something internal to complete before issuing the Operation
Failed error.
Regards,
Glenn Marcy