Il 10/07/2017 13:49, knarra ha scritto:
On 07/10/2017 04:18 PM, Simone Marchioni wrote:
> Il 10/07/2017 09:08, knarra ha scritto:
>> Hi Simone,
>>
>> Can you please let me know what is the version of gdeploy and
>> ansible on your system? Can you check if the path
>> /usr/share/ansible/gdeploy/scripts/grafton-sanity-check.sh exist ?
>> If not, can you edit the generated config file and change the path
>> to "/usr/share/gdeploy/scripts/grafton-sanity-check.sh and see if
>> that works ?
>>
>> You can check the logs in /var/log/messages , or setting
>> log_path in /etc/ansbile/ansible.cfg file.
>>
>> Thanks
>>
>> kasturi.
>>
>
> Hi Kasturi,
>
> thank you for your reply. Here are my versions:
>
> gdeploy-2.0.2-7.noarch
> ansible-2.3.0.0-3.el7.noarch
>
> The file /usr/share/ansible/gdeploy/scripts/grafton-sanity-check.sh
> is missing. For the sake of completeness, the entire directory
> ansible is missing under /usr/share.
>
> In /var/log/messages there is no error message, and I have no
> /etc/ansbile/ansible.cfg config file...
>
> I'm starting to think there are some missing pieces in my
> installation. I installed the following packages:
>
> yum install ovirt-engine
> yum install ovirt-hosted-engine-setup
> yum install ovirt-engine-setup-plugin-live ovirt-live-artwork-gnome
> libgovirt ovirt-live-artwork ovirt-log-collector gdeploy
> cockpit-ovirt-dashboard
>
> and relative dependencies.
>
> Any idea?
Can you check if "/usr/share/gdeploy/scripts/grafton-sanity-check.sh"
is present ? If yes, can you change the path in your generated gdeploy
config file and run again ?
Hi Kasturi,
you're right: the file
/usr/share/gdeploy/scripts/grafton-sanity-check.sh is present. I updated
the path in the gdeploy config file and run Deploy again.
The situation is much better but the Deployment failed again... :-(
Here are the errors:
PLAY [gluster_servers]
*********************************************************
TASK [Run a shell script]
******************************************************
fatal: [ha1.lynx2000.it]: FAILED! => {"failed": true, "msg":
"The
conditional check 'result.rc != 0' failed. The error was: error while
evaluating conditional (result.rc != 0): 'dict object' has no attribute
'rc'"}
fatal: [ha3.lynx2000.it]: FAILED! => {"failed": true, "msg":
"The
conditional check 'result.rc != 0' failed. The error was: error while
evaluating conditional (result.rc != 0): 'dict object' has no attribute
'rc'"}
fatal: [ha2.lynx2000.it]: FAILED! => {"failed": true, "msg":
"The
conditional check 'result.rc != 0' failed. The error was: error while
evaluating conditional (result.rc != 0): 'dict object' has no attribute
'rc'"}
to retry, use: --limit @/tmp/tmpNn6XNG/run-script.retry
PLAY RECAP
*********************************************************************
ha1.lynx2000.it : ok=0 changed=0 unreachable=0 failed=1
ha2.lynx2000.it : ok=0 changed=0 unreachable=0 failed=1
ha3.lynx2000.it : ok=0 changed=0 unreachable=0 failed=1
PLAY [gluster_servers]
*********************************************************
TASK [Clean up filesystem signature]
*******************************************
skipping: [ha2.lynx2000.it] => (item=/dev/md128)
skipping: [ha1.lynx2000.it] => (item=/dev/md128)
skipping: [ha3.lynx2000.it] => (item=/dev/md128)
TASK [Create Physical Volume]
**************************************************
failed: [ha2.lynx2000.it] (item=/dev/md128) => {"failed": true,
"failed_when_result": true, "item": "/dev/md128",
"msg": "WARNING: xfs
signature detected on /dev/md128 at offset 0. Wipe it? [y/n]: [n]\n
Aborted wiping of xfs.\n 1 existing signature left on the device.\n",
"rc": 5}
failed: [ha1.lynx2000.it] (item=/dev/md128) => {"failed": true,
"failed_when_result": true, "item": "/dev/md128",
"msg": "WARNING: xfs
signature detected on /dev/md128 at offset 0. Wipe it? [y/n]: [n]\n
Aborted wiping of xfs.\n 1 existing signature left on the device.\n",
"rc": 5}
failed: [ha3.lynx2000.it] (item=/dev/md128) => {"failed": true,
"failed_when_result": true, "item": "/dev/md128",
"msg": "WARNING: xfs
signature detected on /dev/md128 at offset 0. Wipe it? [y/n]: [n]\n
Aborted wiping of xfs.\n 1 existing signature left on the device.\n",
"rc": 5}
to retry, use: --limit @/tmp/tmpNn6XNG/pvcreate.retry
PLAY RECAP
*********************************************************************
ha1.lynx2000.it : ok=0 changed=0 unreachable=0 failed=1
ha2.lynx2000.it : ok=0 changed=0 unreachable=0 failed=1
ha3.lynx2000.it : ok=0 changed=0 unreachable=0 failed=1
Ignoring errors...
Any clue?
Thanks for your time.
Simone