
Il 10/07/2017 13:49, knarra ha scritto:
On 07/10/2017 04:18 PM, Simone Marchioni wrote:
Il 10/07/2017 09:08, knarra ha scritto:
Hi Simone,
Can you please let me know what is the version of gdeploy and ansible on your system? Can you check if the path /usr/share/ansible/gdeploy/scripts/grafton-sanity-check.sh exist ? If not, can you edit the generated config file and change the path to "/usr/share/gdeploy/scripts/grafton-sanity-check.sh and see if that works ?
You can check the logs in /var/log/messages , or setting log_path in /etc/ansbile/ansible.cfg file.
Thanks
kasturi.
Hi Kasturi,
thank you for your reply. Here are my versions:
gdeploy-2.0.2-7.noarch ansible-2.3.0.0-3.el7.noarch
The file /usr/share/ansible/gdeploy/scripts/grafton-sanity-check.sh is missing. For the sake of completeness, the entire directory ansible is missing under /usr/share.
In /var/log/messages there is no error message, and I have no /etc/ansbile/ansible.cfg config file...
I'm starting to think there are some missing pieces in my installation. I installed the following packages:
yum install ovirt-engine yum install ovirt-hosted-engine-setup yum install ovirt-engine-setup-plugin-live ovirt-live-artwork-gnome libgovirt ovirt-live-artwork ovirt-log-collector gdeploy cockpit-ovirt-dashboard
and relative dependencies.
Any idea? Can you check if "/usr/share/gdeploy/scripts/grafton-sanity-check.sh" is present ? If yes, can you change the path in your generated gdeploy config file and run again ?
Hi Kasturi, you're right: the file /usr/share/gdeploy/scripts/grafton-sanity-check.sh is present. I updated the path in the gdeploy config file and run Deploy again. The situation is much better but the Deployment failed again... :-( Here are the errors: PLAY [gluster_servers] ********************************************************* TASK [Run a shell script] ****************************************************** fatal: [ha1.lynx2000.it]: FAILED! => {"failed": true, "msg": "The conditional check 'result.rc != 0' failed. The error was: error while evaluating conditional (result.rc != 0): 'dict object' has no attribute 'rc'"} fatal: [ha3.lynx2000.it]: FAILED! => {"failed": true, "msg": "The conditional check 'result.rc != 0' failed. The error was: error while evaluating conditional (result.rc != 0): 'dict object' has no attribute 'rc'"} fatal: [ha2.lynx2000.it]: FAILED! => {"failed": true, "msg": "The conditional check 'result.rc != 0' failed. The error was: error while evaluating conditional (result.rc != 0): 'dict object' has no attribute 'rc'"} to retry, use: --limit @/tmp/tmpNn6XNG/run-script.retry PLAY RECAP ********************************************************************* ha1.lynx2000.it : ok=0 changed=0 unreachable=0 failed=1 ha2.lynx2000.it : ok=0 changed=0 unreachable=0 failed=1 ha3.lynx2000.it : ok=0 changed=0 unreachable=0 failed=1 PLAY [gluster_servers] ********************************************************* TASK [Clean up filesystem signature] ******************************************* skipping: [ha2.lynx2000.it] => (item=/dev/md128) skipping: [ha1.lynx2000.it] => (item=/dev/md128) skipping: [ha3.lynx2000.it] => (item=/dev/md128) TASK [Create Physical Volume] ************************************************** failed: [ha2.lynx2000.it] (item=/dev/md128) => {"failed": true, "failed_when_result": true, "item": "/dev/md128", "msg": "WARNING: xfs signature detected on /dev/md128 at offset 0. Wipe it? [y/n]: [n]\n Aborted wiping of xfs.\n 1 existing signature left on the device.\n", "rc": 5} failed: [ha1.lynx2000.it] (item=/dev/md128) => {"failed": true, "failed_when_result": true, "item": "/dev/md128", "msg": "WARNING: xfs signature detected on /dev/md128 at offset 0. Wipe it? [y/n]: [n]\n Aborted wiping of xfs.\n 1 existing signature left on the device.\n", "rc": 5} failed: [ha3.lynx2000.it] (item=/dev/md128) => {"failed": true, "failed_when_result": true, "item": "/dev/md128", "msg": "WARNING: xfs signature detected on /dev/md128 at offset 0. Wipe it? [y/n]: [n]\n Aborted wiping of xfs.\n 1 existing signature left on the device.\n", "rc": 5} to retry, use: --limit @/tmp/tmpNn6XNG/pvcreate.retry PLAY RECAP ********************************************************************* ha1.lynx2000.it : ok=0 changed=0 unreachable=0 failed=1 ha2.lynx2000.it : ok=0 changed=0 unreachable=0 failed=1 ha3.lynx2000.it : ok=0 changed=0 unreachable=0 failed=1 Ignoring errors... Any clue? Thanks for your time. Simone