[ovirt-users] Installation of oVirt 4.1, Gluster Storage and Hosted Engine

Simone Marchioni s.marchioni at lynx2000.it
Mon Jul 10 13:41:34 UTC 2017


Il 10/07/2017 13:06, Gianluca Cecchi ha scritto:
> On Mon, Jul 10, 2017 at 12:57 PM, Simone Marchioni 
> <s.marchioni at lynx2000.it <mailto:s.marchioni at lynx2000.it>> wrote:
>
>
>     Hi Gianluca,
>
>     I recently discovered that the file:
>
>     /usr/share/ansible/gdeploy/scripts/grafton-sanity-check.sh
>
>     is missing from the system, and probably is the root cause of my
>     problem.
>     Searched with
>
>     yum provides
>
>     but I can't find any package with the script inside... any clue?
>
>     Thank you
>     Simone
>
>     _______________________________________________
>     Users mailing list
>     Users at ovirt.org <mailto:Users at ovirt.org>
>     http://lists.ovirt.org/mailman/listinfo/users
>     <http://lists.ovirt.org/mailman/listinfo/users>
>
>
> Hi,
> but are your nodes ovirt-ng nodes or plain CentOS 7.3 where you 
> manually installed packages?
> Becase the original web link covered the case of ovirt-ng nodes, not 
> CentOS 7.3 OS.
> Possibly you are missing any package that is instead installed inside 
> ovirt-ng node by default?
>


Hi Gianluca,

I used plain CentOS 7.3 where I manually installed the necessary packages.
I know the original tutorial used oVirt Node, but I tought the two were 
almost the same, with the latter an "out of the box" solution but with 
the same features.

That said I discovered the problem: there is no missing package. The 
path of the script is wrong. In the tutorial it says:

/usr/share/ansible/gdeploy/scripts/grafton-sanity-check.sh

while the installed script is in:

/usr/share/gdeploy/scripts/grafton-sanity-check.sh

and is (correctly) part of the gdeploy package.

Updated the Gdeploy config and executed Deploy again. The situation is 
much better now, but still says "Deployment Failed". Here's the output:


PLAY [gluster_servers] 
*********************************************************

TASK [Run a shell script] 
******************************************************
changed: [ha3.domain.it] => 
(item=/usr/share/gdeploy/scripts/grafton-sanity-check.sh -d md128 -h 
ha1.domain.it,ha2.domain.it,ha3.domain.it)
changed: [ha2.domain.it] => 
(item=/usr/share/gdeploy/scripts/grafton-sanity-check.sh -d md128 -h 
ha1.domain.it,ha2.domain.it,ha3.domain.it)
changed: [ha1.domain.it] => 
(item=/usr/share/gdeploy/scripts/grafton-sanity-check.sh -d md128 -h 
ha1.domain.it,ha2.domain.it,ha3.domain.it)

PLAY RECAP 
*********************************************************************
ha1.domain.it            : ok=1    changed=1    unreachable=0 failed=0
ha2.domain.it            : ok=1    changed=1    unreachable=0 failed=0
ha3.domain.it            : ok=1    changed=1    unreachable=0 failed=0


PLAY [gluster_servers] 
*********************************************************

TASK [Enable or disable services] 
**********************************************
ok: [ha1.domain.it] => (item=chronyd)
ok: [ha3.domain.it] => (item=chronyd)
ok: [ha2.domain.it] => (item=chronyd)

PLAY RECAP 
*********************************************************************
ha1.lynx2000.it            : ok=1    changed=0    unreachable=0 failed=0
ha2.lynx2000.it            : ok=1    changed=0    unreachable=0 failed=0
ha3.lynx2000.it            : ok=1    changed=0    unreachable=0 failed=0


PLAY [gluster_servers] 
*********************************************************

TASK [start/stop/restart/reload services] 
**************************************
changed: [ha1.domain.it] => (item=chronyd)
changed: [ha2.domain.it] => (item=chronyd)
changed: [ha3.domain.it] => (item=chronyd)

PLAY RECAP 
*********************************************************************
ha1.domain.it            : ok=1    changed=1    unreachable=0 failed=0
ha2.domain.it            : ok=1    changed=1    unreachable=0 failed=0
ha3.domain.it            : ok=1    changed=1    unreachable=0 failed=0


PLAY [gluster_servers] 
*********************************************************

TASK [Run a command in the shell] 
**********************************************
changed: [ha1.domain.it] => (item=vdsm-tool configure --force)
changed: [ha3.domain.it] => (item=vdsm-tool configure --force)
changed: [ha2.domain.it] => (item=vdsm-tool configure --force)

PLAY RECAP 
*********************************************************************
ha1.lynx2000.it            : ok=1    changed=1    unreachable=0 failed=0
ha2.lynx2000.it            : ok=1    changed=1    unreachable=0 failed=0
ha3.lynx2000.it            : ok=1    changed=1    unreachable=0 failed=0


PLAY [gluster_servers] 
*********************************************************

TASK [Run a shell script] 
******************************************************
fatal: [ha1.domain.it]: FAILED! => {"failed": true, "msg": "The 
conditional check 'result.rc != 0' failed. The error was: error while 
evaluating conditional (result.rc != 0): 'dict object' has no attribute 
'rc'"}
fatal: [ha3.domain.it]: FAILED! => {"failed": true, "msg": "The 
conditional check 'result.rc != 0' failed. The error was: error while 
evaluating conditional (result.rc != 0): 'dict object' has no attribute 
'rc'"}
fatal: [ha2.domain.it]: FAILED! => {"failed": true, "msg": "The 
conditional check 'result.rc != 0' failed. The error was: error while 
evaluating conditional (result.rc != 0): 'dict object' has no attribute 
'rc'"}
     to retry, use: --limit @/tmp/tmpNn6XNG/run-script.retry

PLAY RECAP 
*********************************************************************
ha1.lynx2000.it            : ok=0    changed=0    unreachable=0 failed=1
ha2.lynx2000.it            : ok=0    changed=0    unreachable=0 failed=1
ha3.lynx2000.it            : ok=0    changed=0    unreachable=0 failed=1


PLAY [gluster_servers] 
*********************************************************

TASK [Clean up filesystem signature] 
*******************************************
skipping: [ha2.domain.it] => (item=/dev/md128)
skipping: [ha1.domain.it] => (item=/dev/md128)
skipping: [ha3.domain.it] => (item=/dev/md128)

TASK [Create Physical Volume] 
**************************************************
failed: [ha2.domain.it] (item=/dev/md128) => {"failed": true, 
"failed_when_result": true, "item": "/dev/md128", "msg": "WARNING: xfs 
signature detected on /dev/md128 at offset 0. Wipe it? [y/n]: [n]\n  
Aborted wiping of xfs.\n  1 existing signature left on the device.\n", 
"rc": 5}
failed: [ha1.domain.it] (item=/dev/md128) => {"failed": true, 
"failed_when_result": true, "item": "/dev/md128", "msg": "WARNING: xfs 
signature detected on /dev/md128 at offset 0. Wipe it? [y/n]: [n]\n  
Aborted wiping of xfs.\n  1 existing signature left on the device.\n", 
"rc": 5}
failed: [ha3.domain.it] (item=/dev/md128) => {"failed": true, 
"failed_when_result": true, "item": "/dev/md128", "msg": "WARNING: xfs 
signature detected on /dev/md128 at offset 0. Wipe it? [y/n]: [n]\n  
Aborted wiping of xfs.\n  1 existing signature left on the device.\n", 
"rc": 5}
     to retry, use: --limit @/tmp/tmpNn6XNG/pvcreate.retry

PLAY RECAP 
*********************************************************************
ha1.domain.it            : ok=0    changed=0    unreachable=0 failed=1
ha2.domain.it            : ok=0    changed=0    unreachable=0 failed=1
ha3.domain.it            : ok=0    changed=0    unreachable=0 failed=1

Ignoring errors...


Hope to be near the solution... ;-)

Hi,
Simone
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20170710/87549159/attachment.html>


More information about the Users mailing list