[ovirt-users] Installation of oVirt 4.1, Gluster Storage and Hosted Engine
Simone Marchioni
s.marchioni at lynx2000.it
Mon Jul 10 13:48:36 UTC 2017
Il 10/07/2017 13:49, knarra ha scritto:
> On 07/10/2017 04:18 PM, Simone Marchioni wrote:
>> Il 10/07/2017 09:08, knarra ha scritto:
>>> Hi Simone,
>>>
>>> Can you please let me know what is the version of gdeploy and
>>> ansible on your system? Can you check if the path
>>> /usr/share/ansible/gdeploy/scripts/grafton-sanity-check.sh exist ?
>>> If not, can you edit the generated config file and change the path
>>> to "/usr/share/gdeploy/scripts/grafton-sanity-check.sh and see if
>>> that works ?
>>>
>>> You can check the logs in /var/log/messages , or setting
>>> log_path in /etc/ansbile/ansible.cfg file.
>>>
>>> Thanks
>>>
>>> kasturi.
>>>
>>
>> Hi Kasturi,
>>
>> thank you for your reply. Here are my versions:
>>
>> gdeploy-2.0.2-7.noarch
>> ansible-2.3.0.0-3.el7.noarch
>>
>> The file /usr/share/ansible/gdeploy/scripts/grafton-sanity-check.sh
>> is missing. For the sake of completeness, the entire directory
>> ansible is missing under /usr/share.
>>
>> In /var/log/messages there is no error message, and I have no
>> /etc/ansbile/ansible.cfg config file...
>>
>> I'm starting to think there are some missing pieces in my
>> installation. I installed the following packages:
>>
>> yum install ovirt-engine
>> yum install ovirt-hosted-engine-setup
>> yum install ovirt-engine-setup-plugin-live ovirt-live-artwork-gnome
>> libgovirt ovirt-live-artwork ovirt-log-collector gdeploy
>> cockpit-ovirt-dashboard
>>
>> and relative dependencies.
>>
>> Any idea?
> Can you check if "/usr/share/gdeploy/scripts/grafton-sanity-check.sh"
> is present ? If yes, can you change the path in your generated gdeploy
> config file and run again ?
Hi Kasturi,
you're right: the file
/usr/share/gdeploy/scripts/grafton-sanity-check.sh is present. I updated
the path in the gdeploy config file and run Deploy again.
The situation is much better but the Deployment failed again... :-(
Here are the errors:
PLAY [gluster_servers]
*********************************************************
TASK [Run a shell script]
******************************************************
fatal: [ha1.lynx2000.it]: FAILED! => {"failed": true, "msg": "The
conditional check 'result.rc != 0' failed. The error was: error while
evaluating conditional (result.rc != 0): 'dict object' has no attribute
'rc'"}
fatal: [ha3.lynx2000.it]: FAILED! => {"failed": true, "msg": "The
conditional check 'result.rc != 0' failed. The error was: error while
evaluating conditional (result.rc != 0): 'dict object' has no attribute
'rc'"}
fatal: [ha2.lynx2000.it]: FAILED! => {"failed": true, "msg": "The
conditional check 'result.rc != 0' failed. The error was: error while
evaluating conditional (result.rc != 0): 'dict object' has no attribute
'rc'"}
to retry, use: --limit @/tmp/tmpNn6XNG/run-script.retry
PLAY RECAP
*********************************************************************
ha1.lynx2000.it : ok=0 changed=0 unreachable=0 failed=1
ha2.lynx2000.it : ok=0 changed=0 unreachable=0 failed=1
ha3.lynx2000.it : ok=0 changed=0 unreachable=0 failed=1
PLAY [gluster_servers]
*********************************************************
TASK [Clean up filesystem signature]
*******************************************
skipping: [ha2.lynx2000.it] => (item=/dev/md128)
skipping: [ha1.lynx2000.it] => (item=/dev/md128)
skipping: [ha3.lynx2000.it] => (item=/dev/md128)
TASK [Create Physical Volume]
**************************************************
failed: [ha2.lynx2000.it] (item=/dev/md128) => {"failed": true,
"failed_when_result": true, "item": "/dev/md128", "msg": "WARNING: xfs
signature detected on /dev/md128 at offset 0. Wipe it? [y/n]: [n]\n
Aborted wiping of xfs.\n 1 existing signature left on the device.\n",
"rc": 5}
failed: [ha1.lynx2000.it] (item=/dev/md128) => {"failed": true,
"failed_when_result": true, "item": "/dev/md128", "msg": "WARNING: xfs
signature detected on /dev/md128 at offset 0. Wipe it? [y/n]: [n]\n
Aborted wiping of xfs.\n 1 existing signature left on the device.\n",
"rc": 5}
failed: [ha3.lynx2000.it] (item=/dev/md128) => {"failed": true,
"failed_when_result": true, "item": "/dev/md128", "msg": "WARNING: xfs
signature detected on /dev/md128 at offset 0. Wipe it? [y/n]: [n]\n
Aborted wiping of xfs.\n 1 existing signature left on the device.\n",
"rc": 5}
to retry, use: --limit @/tmp/tmpNn6XNG/pvcreate.retry
PLAY RECAP
*********************************************************************
ha1.lynx2000.it : ok=0 changed=0 unreachable=0 failed=1
ha2.lynx2000.it : ok=0 changed=0 unreachable=0 failed=1
ha3.lynx2000.it : ok=0 changed=0 unreachable=0 failed=1
Ignoring errors...
Any clue?
Thanks for your time.
Simone
More information about the Users
mailing list