And there it is the last lines on the ansible_create_storage_domain log:

2019-04-02 10:53:49,139+0100 DEBUG var changed: host "localhost" var "otopi_storage_domain_details" type "<type 'dict'>" value: "{
    "changed": false,
    "exception": "Traceback (most recent call last):\n  File \"/tmp/ansible_ovirt_storage_domain_payload_6Jxg5v/__main__.py\", line 664, in main\n    storage_domains_module.post_create_check(sd_id)\n  File \"/tmp/ansible_ovirt_storage_domain_payload_6Jxg5v/__main__.py\", line 526, in post_create_check\n    id=storage_domain.id,\n  File \"/usr/lib64/python2.7/site-packages/ovirtsdk4/services.py\", line 3053, in add\n    return self._internal_add(storage_domain, headers, query, wait)\n  File \"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 232, in _internal_add\n    return future.wait() if wait else future\n  File \"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 55, in wait\n    return self._code(response)\n  File \"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 229, in callback\n    self._check_fault(response)\n  File \"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 132, in _check_fault\n    self._raise_error(response, body)\n  File \"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 118, in _raise_error\n    raise error\nError: Fault reason is \"Operation Failed\". Fault detail is \"[]\". HTTP response code is 400.\n",
    "failed": true,
    "msg": "Fault reason is \"Operation Failed\". Fault detail is \"[]\". HTTP response code is 400."
}"
2019-04-02 10:53:49,141+0100 DEBUG var changed: host "localhost" var "ansible_play_hosts" type "<type 'list'>" value: "[]"
2019-04-02 10:53:49,141+0100 DEBUG var changed: host "localhost" var "play_hosts" type "<type 'list'>" value: "[]"
2019-04-02 10:53:49,142+0100 DEBUG var changed: host "localhost" var "ansible_play_batch" type "<type 'list'>" value: "[]"
2019-04-02 10:53:49,142+0100 ERROR ansible failed {'status': 'FAILED', 'ansible_type': 'task', 'ansible_task': u'Activate storage domain', 'ansible_result': u'type: <type \'dict\'>\nstr: {\'_ansible_parsed\': True, u\'exception\': u\'Traceback (most recent call last):\\n  File "/tmp/ansible_ovirt_storage_domain_payload_6Jxg5v/__main__.py", line 664, in main\\n    storage_domains_module.post_create_check(sd_id)\\n  File "/tmp/ansible_ovirt_storage_domain_payload_6Jxg5v/__main__.py", line 526', 'task_duration': 9, 'ansible_host': u'localhost', 'ansible_playbook': u'/usr/share/ovirt-hosted-engine-setup/ansible/trigger_role.yml'}
2019-04-02 10:53:49,143+0100 DEBUG ansible on_any args <ansible.executor.task_result.TaskResult object at 0x7f03fd025e50> kwargs ignore_errors:None
2019-04-02 10:53:49,148+0100 INFO ansible stats {
    "ansible_playbook": "/usr/share/ovirt-hosted-engine-setup/ansible/trigger_role.yml",
    "ansible_playbook_duration": "01:15 Minutes",
    "ansible_result": "type: <type 'dict'>\nstr: {u'localhost': {'unreachable': 0, 'skipped': 6, 'ok': 23, 'changed': 0, 'failures': 1}}",
    "ansible_type": "finish",
    "status": "FAILED"
}
2019-04-02 10:53:49,149+0100 INFO SUMMARY:
Duration        Task Name
--------        --------
[ < 1 sec ]     Execute just a specific set of steps
[  00:02  ]     Force facts gathering
[  00:02  ]     Check local VM dir stat
[  00:02  ]     Obtain SSO token using username/password credentials
[  00:02  ]     Fetch host facts
[  00:01  ]     Fetch cluster ID
[  00:02  ]     Fetch cluster facts
[  00:02  ]     Fetch Datacenter facts
[  00:01  ]     Fetch Datacenter ID
[  00:01  ]     Fetch Datacenter name
[  00:02  ]     Add glusterfs storage domain
[  00:02  ]     Get storage domain details
[  00:02  ]     Find the appliance OVF
[  00:02  ]     Parse OVF
[  00:02  ]     Get required size
[ FAILED  ]     Activate storage domain

Any ideea on how to escalate this issue ?
It just does not make sense to not be able to install from scratch a fresh node...

Have a nice day  !

Leo


On Tue, Apr 2, 2019 at 12:11 PM Gobinda Das <godas@redhat.com> wrote:
Hi Leo,
 Can you please paste "df -Th" and "gluster v status" out put ?
Wanted to make sure engine mounted and volumes and bricks are up.
What does vdsm log say?

On Tue, Apr 2, 2019 at 2:06 PM Leo David <leoalex@gmail.com> wrote:
Thank you very much !
I have just installed a new fresh node,   and triggered the single instance hyperconverged setup. It seems it fails at the hosted engine final steps of deployment:
 INFO ] TASK [ovirt.hosted_engine_setup : Get required size]
[ INFO ] ok: [localhost]
[ INFO ] TASK [ovirt.hosted_engine_setup : Remove unsuitable storage domain]
[ INFO ] skipping: [localhost]
[ INFO ] TASK [ovirt.hosted_engine_setup : Check storage domain free space]
[ INFO ] skipping: [localhost]
[ INFO ] TASK [ovirt.hosted_engine_setup : Activate storage domain]
[ ERROR ] Error: Fault reason is "Operation Failed". Fault detail is "[Cannot attach Storage. There is no active Host in the Data Center.]". HTTP response code is 409.
[ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": "Fault reason is \"Operation Failed\". Fault detail is \"[Cannot attach Storage. There is no active Host in the Data Center.]\". HTTP response code is 409."}
Also,  the ovirt-hosted-engine-setup-ansible-create_storage_domain-201932112413-xkq6nb.log throws the following:

2019-04-02 09:25:40,420+0100 DEBUG var changed: host "localhost" var "otopi_storage_domain_details" type "<type 'dict'>" value: "{
    "changed": false,
    "exception": "Traceback (most recent call last):\n  File \"/tmp/ansible_ovirt_storage_domain_payload_87MSyY/__main__.py\", line 664, in main\n    storage_domains_module.post_create_check(sd_id)\n  File \"/tmp/ansible_ovirt_storage_domain_payload_87MSyY/__main__.py\", line 526, in post_create_check\n    id=storage_domain.id,\n  File \"/usr/lib64/python2.7/site-packages/ovirtsdk4/services.py\", line 3053, in add\n    return self._internal_add(storage_domain, headers, query, wait)\n  File \"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 232, in _internal_add\n    return future.wait() if wait else future\n  File \"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 55, in wait\n    return self._code(response)\n  File \"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 229, in callback\n    self._check_fault(response)\n  File \"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 132, in _check_fault\n    self._raise_error(response, body)\n  File \"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 118, in _raise_error\n    raise error\nError: Fault reason is \"Operation Failed\". Fault detail is \"[Cannot attach Storage. There is no active Host in the Data Center.]\". HTTP response code is 409.\n",
    "failed": true,
    "msg": "Fault reason is \"Operation Failed\". Fault detail is \"[Cannot attach Storage. There is no active Host in the Data Center.]\". HTTP response code is 409."
}"

I have used the ovirt-node-ng-installer-4.3.2-2019031908.el7.iso. So far,  I am unable to deploy oVirt single node Hyperconverged...
Any thoughts ?



On Mon, Apr 1, 2019 at 9:46 PM Simone Tiraboschi <stirabos@redhat.com> wrote:


On Mon, Apr 1, 2019 at 6:14 PM Leo David <leoalex@gmail.com> wrote:
Thank you Simone.
I've decides to go for a new fresh install from iso, and i'll keep posted if any troubles arise. But I am still trying to understand what are the services that mount the lvms and volumes after configuration. There is nothing related in fstab, so I assume there are a couple of .mount files somewhere in the filesystem.
Im just trying to understand node's underneath workflow.

hosted-engine configuration is stored in /etc/ovirt-hosted-engine/hosted-engine.conf; ovirt-ha-broker will mount the hosted-engine storage domain according to that and so ovirt-ha-agent will be able to start the engine VM.
Everything else is just in the engine DB.
 

On Mon, Apr 1, 2019, 10:16 Simone Tiraboschi <stirabos@redhat.com> wrote:
Hi,
to understand what's failing I'd suggest to start attaching setup logs. 

On Sun, Mar 31, 2019 at 5:06 PM Leo David <leoalex@gmail.com> wrote:
Hello Everyone,
Using 4.3.2 installation, and after running through HyperConverged Setup,  at the last stage it fails. It seems that the previously created "engine" volume is not mounted under "/rhev" path, therefore the setup cannot finish the deployment.
Any ideea which are the services responsible of mounting the volumes on oVirt Node distribution ? I'm thinking that maybe this particularly one failed to start for some reason...
Thank you very much !

--
Best regards, Leo David
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-leave@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/
List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/PUXDAQHVNZWF4TIXZ3GIBZHSJ7IC2VHC/


--
Best regards, Leo David
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-leave@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/
List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/NROCMKIFJDEIWD7PELNTBKDQMPMFLDRN/


--


Thanks,
Gobinda


--
Best regards, Leo David