Hi Gobinda,
gdeploy --version
gdeploy 2.0.2
yum list installed | grep gdeploy
gdeploy.noarch 2.0.8-1.el7
installed
Thank you !
On Mon, Jan 28, 2019 at 10:56 AM Gobinda Das <godas(a)redhat.com> wrote:
Hi David,
Can you please check the gdeploy version?
This bug was fixed last year:
https://bugzilla.redhat.com/show_bug.cgi?id=1626513
And is part of: gdeploy-2.0.2-29
On Sun, Jan 27, 2019 at 2:38 PM Leo David <leoalex(a)gmail.com> wrote:
> Hi,
> It seems so that I had to manually add the sections, to make the scrip
> working:
> [diskcount]
> 12
> [stripesize]
> 256
>
> It looks like ansible is still searching for these sections regardless
> that I have configured "jbod" in the wizard...
>
> Thanks,
>
> Leo
>
>
>
> On Sun, Jan 27, 2019 at 10:49 AM Leo David <leoalex(a)gmail.com> wrote:
>
>> Hello Everyone,
>> Using version 4.2.8, ( ovirt-node-ng-installer-4.2.0-2019012606.el7.iso
>> ) for deploying one node instance by following from within CockpitUI seems
>> not to be possible.
>> Here's the generated inventory ( i've specified "jbod" in the
wizard ):
>>
>> #gdeploy configuration generated by cockpit-gluster plugin
>> [hosts]
>> 192.168.80.191
>>
>> [script1:192.168.80.191]
>> action=execute
>> ignore_script_errors=no
>> file=/usr/share/gdeploy/scripts/grafton-sanity-check.sh -d sdb -h
>> 192.168.80.191
>> [disktype]
>> jbod
>> [service1]
>> action=enable
>> service=chronyd
>> [service2]
>> action=restart
>> service=chronyd
>> [shell2]
>> action=execute
>> command=vdsm-tool configure --force
>> [script3]
>> action=execute
>> file=/usr/share/gdeploy/scripts/blacklist_all_disks.sh
>> ignore_script_errors=no
>> [pv1:192.168.80.191]
>> action=create
>> devices=sdb
>> ignore_pv_errors=no
>> [vg1:192.168.80.191]
>> action=create
>> vgname=gluster_vg_sdb
>> pvname=sdb
>> ignore_vg_errors=no
>> [lv1:192.168.80.191]
>> action=create
>> lvname=gluster_lv_engine
>> ignore_lv_errors=no
>> vgname=gluster_vg_sdb
>> mount=/gluster_bricks/engine
>> size=230GB
>> lvtype=thick
>> [selinux]
>> yes
>> [service3]
>> action=restart
>> service=glusterd
>> slice_setup=yes
>> [firewalld]
>> action=add
>>
>>
ports=111/tcp,2049/tcp,54321/tcp,5900/tcp,5900-6923/tcp,5666/tcp,16514/tcp,54322/tcp
>> services=glusterfs
>> [script2]
>> action=execute
>> file=/usr/share/gdeploy/scripts/disable-gluster-hooks.sh
>> [shell3]
>> action=execute
>> command=usermod -a -G gluster qemu
>> [volume1]
>> action=create
>> volname=engine
>> transport=tcp
>>
>>
key=storage.owner-uid,storage.owner-gid,features.shard,performance.low-prio-threads,performance.strict-o-direct,network.remote-dio,network.ping-timeout,user.cifs,nfs.disable,performance.quick-read,performance.read-ahead,performance.io-cache,cluster.eager-lock
>> value=36,36,on,32,on,off,30,off,on,off,off,off,enable
>> brick_dirs=192.168.80.191:/gluster_bricks/engine/engine
>> ignore_volume_errors=no
>>
>> It does not get to finish, throwing the following error:
>>
>> PLAY [gluster_servers]
>> *********************************************************
>> TASK [Create volume group on the disks]
>> ****************************************
>> changed: [192.168.80.191] => (item={u'brick': u'/dev/sdb',
u'vg':
>> u'gluster_vg_sdb'})
>> PLAY RECAP
>> *********************************************************************
>> 192.168.80.191 : ok=1 changed=1 unreachable=0
>> failed=0
>> *Error: Section diskcount not found in the configuration file*
>>
>> Any thoughts ?
>>
>>
>>
>>
>>
>>
>> --
>> Best regards, Leo David
>>
>
>
> --
> Best regards, Leo David
> _______________________________________________
> Users mailing list -- users(a)ovirt.org
> To unsubscribe send an email to users-leave(a)ovirt.org
> Privacy Statement:
https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct:
>
https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
>
https://lists.ovirt.org/archives/list/users@ovirt.org/message/Z2X45A6V6WQ...
>
--
Thanks,
Gobinda