
Now that I've been able to get past my issues with q35 bios I decided to try the hyperconverged install. setting up the storage was no problem and the engine setup went fine until the point where it wanted to use that glusterfs storage. I got the error, pretty quickly I might add, of 2020-06-30 02:17:50,689-0400 DEBUG ansible on_any args <ansible.executor.task_result.TaskResult object at 0x7f83f85b4278> kwargs 2020-06-30 02:17:51,428-0400 INFO ansible task start {'status': 'OK', 'ansible_type': 'task', 'ansible_playbook': '/usr/share/ovirt-hosted-engine-setup/ansible/trigger_role.yml', 'ansible_task': 'ovirt.hosted_engine_setup : Add glusterfs storage domain'} 2020-06-30 02:17:51,428-0400 DEBUG ansible on_any args TASK: ovirt.hosted_engine_setup : Add glusterfs storage domain kwargs is_conditional:False 2020-06-30 02:17:51,429-0400 DEBUG ansible on_any args localhostTASK: ovirt.hosted_engine_setup : Add glusterfs storage domain kwargs 2020-06-30 02:17:53,656-0400 DEBUG var changed: host "localhost" var "otopi_storage_domain_details_gluster" type "<class 'dict'>" value: "{ "changed": false, "exception": "Traceback (most recent call last): File \"/tmp/ansible_ovirt_storage_domain_payload_jh42k_ip/ansible_ovirt_storage_domain_payload.zip/ansible/modules/cloud/ovirt/ovirt_storage_domain.py\", line 792, in main File \"/tmp/ansible_ovirt_storage_domain_payload_jh42k_ip/ansible_ovirt_storage_domain_payload.zip/ansible/module_utils/ovirt.py\", line 623, in create **kwargs File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/services.py\", line 26097, in add return self._internal_add(storage_domain, headers, query, wait) File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 232, in _internal_add return future.wait() if wait else future File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 55, in wait return self._code(response) File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 229, in callback self._check_fault(response) File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 132, in _check_fault self._raise_error(response, body) File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 118, in _raise_error raise error\novirtsdk4.Error: Fault reason is \"Operation Failed\". Fault detail is \"[Failed to fetch Gluster Volume List]\". HTTP response code is 400.\n", "failed": true, "msg": "Fault reason is \"Operation Failed\". Fault detail is \"[Failed to fetch Gluster Volume List]\". HTTP response code is 400." }" The glusterfs cli worked fine by that point, had no issues with volume list, so I went into the node:6900/ovirt-engine forwarder to the appliance and could see that at the time of that error in the event logs there was a message that it was starting to update things and a few seconds later a message that the volumes were available. I am thinking that there is need for a retry loop in this step or that something in the path isn't waiting for something internal to complete before issuing the Operation Failed error. Regards, Glenn Marcy