Is it possible you have not cleared the gluster volume between installs?
What's the corresponding error in vdsm.log?
On Tue, Apr 2, 2019 at 4:07 PM Leo David <leoalex(a)gmail.com> wrote:
>
> And there it is the last lines on the ansible_create_storage_domain log:
>
> 2019-04-02 10:53:49,139+0100 DEBUG var changed: host "localhost" var
"otopi_storage_domain_details" type "<type 'dict'>"
value: "{
> "changed": false,
> "exception": "Traceback (most recent call last):\n File
\"/tmp/ansible_ovirt_storage_domain_payload_6Jxg5v/__main__.py\", line 664, in
main\n storage_domains_module.post_create_check(sd_id)\n File
\"/tmp/ansible_ovirt_storage_domain_payload_6Jxg5v/__main__.py\", line 526, in
post_create_check\n id=storage_domain.id,\n File
\"/usr/lib64/python2.7/site-packages/ovirtsdk4/services.py\", line 3053, in
add\n return self._internal_add(storage_domain, headers, query, wait)\n File
\"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 232, in
_internal_add\n return future.wait() if wait else future\n File
\"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 55, in wait\n
return self._code(response)\n File
\"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 229, in
callback\n self._check_fault(response)\n File
\"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 132, in
_check_fault\n self._raise_error(response, body)\n File
\"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 118, in
_raise_error\n raise error\nError: Fault reason is \"Operation Failed\".
Fault detail is \"[]\". HTTP response code is 400.\n",
> "failed": true,
> "msg": "Fault reason is \"Operation Failed\". Fault
detail is \"[]\". HTTP response code is 400."
> }"
> 2019-04-02 10:53:49,141+0100 DEBUG var changed: host "localhost" var
"ansible_play_hosts" type "<type 'list'>" value:
"[]"
> 2019-04-02 10:53:49,141+0100 DEBUG var changed: host "localhost" var
"play_hosts" type "<type 'list'>" value: "[]"
> 2019-04-02 10:53:49,142+0100 DEBUG var changed: host "localhost" var
"ansible_play_batch" type "<type 'list'>" value:
"[]"
> 2019-04-02 10:53:49,142+0100 ERROR ansible failed {'status':
'FAILED', 'ansible_type': 'task', 'ansible_task':
u'Activate storage domain', 'ansible_result': u'type: <type
\'dict\'>\nstr: {\'_ansible_parsed\': True, u\'exception\':
u\'Traceback (most recent call last):\\n File
"/tmp/ansible_ovirt_storage_domain_payload_6Jxg5v/__main__.py", line 664, in
main\\n storage_domains_module.post_create_check(sd_id)\\n File
"/tmp/ansible_ovirt_storage_domain_payload_6Jxg5v/__main__.py", line 526',
'task_duration': 9, 'ansible_host': u'localhost',
'ansible_playbook':
u'/usr/share/ovirt-hosted-engine-setup/ansible/trigger_role.yml'}
> 2019-04-02 10:53:49,143+0100 DEBUG ansible on_any args
<ansible.executor.task_result.TaskResult object at 0x7f03fd025e50> kwargs
ignore_errors:None
> 2019-04-02 10:53:49,148+0100 INFO ansible stats {
> "ansible_playbook":
"/usr/share/ovirt-hosted-engine-setup/ansible/trigger_role.yml",
> "ansible_playbook_duration": "01:15 Minutes",
> "ansible_result": "type: <type 'dict'>\nstr:
{u'localhost': {'unreachable': 0, 'skipped': 6, 'ok': 23,
'changed': 0, 'failures': 1}}",
> "ansible_type": "finish",
> "status": "FAILED"
> }
> 2019-04-02 10:53:49,149+0100 INFO SUMMARY:
> Duration Task Name
> -------- --------
> [ < 1 sec ] Execute just a specific set of steps
> [ 00:02 ] Force facts gathering
> [ 00:02 ] Check local VM dir stat
> [ 00:02 ] Obtain SSO token using username/password credentials
> [ 00:02 ] Fetch host facts
> [ 00:01 ] Fetch cluster ID
> [ 00:02 ] Fetch cluster facts
> [ 00:02 ] Fetch Datacenter facts
> [ 00:01 ] Fetch Datacenter ID
> [ 00:01 ] Fetch Datacenter name
> [ 00:02 ] Add glusterfs storage domain
> [ 00:02 ] Get storage domain details
> [ 00:02 ] Find the appliance OVF
> [ 00:02 ] Parse OVF
> [ 00:02 ] Get required size
> [ FAILED ] Activate storage domain
>
> Any ideea on how to escalate this issue ?
> It just does not make sense to not be able to install from scratch a fresh node...
>
> Have a nice day !
>
> Leo
>
>
> On Tue, Apr 2, 2019 at 12:11 PM Gobinda Das <godas(a)redhat.com> wrote:
>>
>> Hi Leo,
>> Can you please paste "df -Th" and "gluster v status" out put
?
>> Wanted to make sure engine mounted and volumes and bricks are up.
>> What does vdsm log say?
>>
>> On Tue, Apr 2, 2019 at 2:06 PM Leo David <leoalex(a)gmail.com> wrote:
>>>
>>> Thank you very much !
>>> I have just installed a new fresh node, and triggered the single instance
hyperconverged setup. It seems it fails at the hosted engine final steps of deployment:
>>> INFO ] TASK [ovirt.hosted_engine_setup : Get required size]
>>> [ INFO ] ok: [localhost]
>>> [ INFO ] TASK [ovirt.hosted_engine_setup : Remove unsuitable storage domain]
>>> [ INFO ] skipping: [localhost]
>>> [ INFO ] TASK [ovirt.hosted_engine_setup : Check storage domain free space]
>>> [ INFO ] skipping: [localhost]
>>> [ INFO ] TASK [ovirt.hosted_engine_setup : Activate storage domain]
>>> [ ERROR ] Error: Fault reason is "Operation Failed". Fault detail
is "[Cannot attach Storage. There is no active Host in the Data Center.]". HTTP
response code is 409.
>>> [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false,
"msg": "Fault reason is \"Operation Failed\". Fault detail is
\"[Cannot attach Storage. There is no active Host in the Data Center.]\". HTTP
response code is 409."}
>>> Also, the
ovirt-hosted-engine-setup-ansible-create_storage_domain-201932112413-xkq6nb.log throws the
following:
>>>
>>> 2019-04-02 09:25:40,420+0100 DEBUG var changed: host "localhost"
var "otopi_storage_domain_details" type "<type 'dict'>"
value: "{
>>> "changed": false,
>>> "exception": "Traceback (most recent call last):\n File
\"/tmp/ansible_ovirt_storage_domain_payload_87MSyY/__main__.py\", line 664, in
main\n storage_domains_module.post_create_check(sd_id)\n File
\"/tmp/ansible_ovirt_storage_domain_payload_87MSyY/__main__.py\", line 526, in
post_create_check\n id=storage_domain.id,\n File
\"/usr/lib64/python2.7/site-packages/ovirtsdk4/services.py\", line 3053, in
add\n return self._internal_add(storage_domain, headers, query, wait)\n File
\"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 232, in
_internal_add\n return future.wait() if wait else future\n File
\"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 55, in wait\n
return self._code(response)\n File
\"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 229, in
callback\n self._check_fault(response)\n File
\"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 132, in
_check_fault\n self._raise_error(response, body)\n File
\"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 118, in
_raise_error\n raise error\nError: Fault reason is \"Operation Failed\".
Fault detail is \"[Cannot attach Storage. There is no active Host in the Data
Center.]\". HTTP response code is 409.\n",
>>> "failed": true,
>>> "msg": "Fault reason is \"Operation Failed\".
Fault detail is \"[Cannot attach Storage. There is no active Host in the Data
Center.]\". HTTP response code is 409."
>>> }"
>>>
>>> I have used the ovirt-node-ng-installer-4.3.2-2019031908.el7.iso. So far, I
am unable to deploy oVirt single node Hyperconverged...
>>> Any thoughts ?
>>>
>>>
>>>
>>> On Mon, Apr 1, 2019 at 9:46 PM Simone Tiraboschi <stirabos(a)redhat.com>
wrote:
>>>>
>>>>
>>>>
>>>> On Mon, Apr 1, 2019 at 6:14 PM Leo David <leoalex(a)gmail.com>
wrote:
>>>>>
>>>>> Thank you Simone.
>>>>> I've decides to go for a new fresh install from iso, and i'll
keep posted if any troubles arise. But I am still trying to understand what are the
services that mount the lvms and volumes after configuration. There is nothing related in
fstab, so I assume there are a couple of .mount files somewhere in the filesystem.
>>>>> Im just trying to understand node's underneath workflow.
>>>>
>>>>
>>>> hosted-engine configuration is stored in
/etc/ovirt-hosted-engine/hosted-engine.conf; ovirt-ha-broker will mount the hosted-engine
storage domain according to that and so ovirt-ha-agent will be able to start the engine
VM.
>>>> Everything else is just in the engine DB.
>>>>
>>>>>
>>>>>
>>>>> On Mon, Apr 1, 2019, 10:16 Simone Tiraboschi
<stirabos(a)redhat.com> wrote:
>>>>>>
>>>>>> Hi,
>>>>>> to understand what's failing I'd suggest to start
attaching setup logs.
>>>>>>
>>>>>> On Sun, Mar 31, 2019 at 5:06 PM Leo David
<leoalex(a)gmail.com> wrote:
>>>>>>>
>>>>>>> Hello Everyone,
>>>>>>> Using 4.3.2 installation, and after running through
HyperConverged Setup, at the last stage it fails. It seems that the previously created
"engine" volume is not mounted under "/rhev" path, therefore the setup
cannot finish the deployment.
>>>>>>> Any ideea which are the services responsible of mounting the
volumes on oVirt Node distribution ? I'm thinking that maybe this particularly one
failed to start for some reason...
>>>>>>> Thank you very much !
>>>>>>>
>>>>>>> --
>>>>>>> Best regards, Leo David
>>>>>>> _______________________________________________
>>>>>>> Users mailing list -- users(a)ovirt.org
>>>>>>> To unsubscribe send an email to users-leave(a)ovirt.org
>>>>>>> Privacy Statement:
https://www.ovirt.org/site/privacy-policy/
>>>>>>> oVirt Code of Conduct:
https://www.ovirt.org/community/about/community-guidelines/
>>>>>>> List Archives:
https://lists.ovirt.org/archives/list/users@ovirt.org/message/PUXDAQHVNZW...
>>>
>>>
>>>
>>> --
>>> Best regards, Leo David
>>> _______________________________________________
>>> Users mailing list -- users(a)ovirt.org
>>> To unsubscribe send an email to users-leave(a)ovirt.org
>>> Privacy Statement:
https://www.ovirt.org/site/privacy-policy/
>>> oVirt Code of Conduct:
https://www.ovirt.org/community/about/community-guidelines/
>>> List Archives:
https://lists.ovirt.org/archives/list/users@ovirt.org/message/NROCMKIFJDE...
>>
>>
>>
>> --
>>
>>
>> Thanks,
>> Gobinda
>
>
>
> --
> Best regards, Leo David
> _______________________________________________
> Users mailing list -- users(a)ovirt.org
> To unsubscribe send an email to users-leave(a)ovirt.org
> Privacy Statement:
https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct:
https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
https://lists.ovirt.org/archives/list/users@ovirt.org/message/XDJNBS6EOXC...