I got the Gluster working but I can’t finish setting up the engine it fails and says can’t query dns for engine and then can’t remove the storage domain without redoing everything all over again. 



On Tue, Feb 26, 2019 at 2:27 AM Parth Dhanjal <dparth@redhat.com> wrote:
Hey Matthew!

Can you please provide me with the following to help you debug the issue that you are facing?
1. oVirt and gdeploy version
2. /var/log/messages file
3. /root/.gdeploy file

On Mon, Feb 25, 2019 at 1:23 PM Parth Dhanjal <dparth@redhat.com> wrote:
Hey Matthew!

Can you please provide which oVirt and gdeploy version have you installed?

Regards
Parth Dhanjal

On Mon, Feb 25, 2019 at 12:56 PM Sahina Bose <sabose@redhat.com> wrote:
+Gobinda Das +Dhanjal Parth can you please check?

On Fri, Feb 22, 2019 at 11:52 PM Matthew Roth <mroth381@gmail.com> wrote:
>
> I have 3 servers,  Node 1 is 3tb /dev/sda, Node 2, 3tb /dev/sdb,  node3 3tb /dev/sdb
>
> I start the process for gluster deployment. I change node 1 to sda and all the other ones to sdb. I get no errors however,
>
> when I get to
> Creating physical Volume all it does is spin forever . doesnt get any further. I can leave it there for 5 hours and doesn't go anywhere.
>
> #gdeploy configuration generated by cockpit-gluster plugin
> [hosts]
> cmdnode1.cmd911.com
> cmdnode2.cmd911.com
> cmdnode3.cmd911.com
>
> [script1:cmdnode1.cmd911.com]
> action=execute
> ignore_script_errors=no
> file=/usr/share/gdeploy/scripts/grafton-sanity-check.sh -d sda -h cmdnode1.cmd911.com, cmdnode2.cmd911.com, cmdnode3.cmd911.com
>
> [script1:cmdnode2.cmd911.com]
> action=execute
> ignore_script_errors=no
> file=/usr/share/gdeploy/scripts/grafton-sanity-check.sh -d sdb -h cmdnode1.cmd911.com, cmdnode2.cmd911.com, cmdnode3.cmd911.com
>
> [script1:cmdnode3.cmd911.com]
> action=execute
> ignore_script_errors=no
> file=/usr/share/gdeploy/scripts/grafton-sanity-check.sh -d sdb -h cmdnode1.cmd911.com, cmdnode2.cmd911.com, cmdnode3.cmd911.com
>
> [disktype]
> raid6
>
> [diskcount]
> 12
>
> [stripesize]
> 256
>
> [service1]
> action=enable
> service=chronyd
>
> [service2]
> action=restart
> service=chronyd
>
> [shell2]
> action=execute
> command=vdsm-tool configure --force
>
> [script3]
> action=execute
> file=/usr/share/gdeploy/scripts/blacklist_all_disks.sh
> ignore_script_errors=no
>
> [pv1:cmdnode1.cmd911.com]
> action=create
> devices=sda
> ignore_pv_errors=no
>
> [pv1:cmdnode2.cmd911.com]
> action=create
> devices=sdb
> ignore_pv_errors=no
>
> [pv1:cmdnode3.cmd911.com]
> action=create
> devices=sdb
> ignore_pv_errors=no
>
> [vg1:cmdnode1.cmd911.com]
> action=create
> vgname=gluster_vg_sda
> pvname=sda
> ignore_vg_errors=no
>
> [vg1:cmdnode2.cmd911.com]
> action=create
> vgname=gluster_vg_sdb
> pvname=sdb
> ignore_vg_errors=no
>
> [vg1:cmdnode3.cmd911.com]
> action=create
> vgname=gluster_vg_sdb
> pvname=sdb
> ignore_vg_errors=no
>
> [lv1:cmdnode1.cmd911.com]
> action=create
> poolname=gluster_thinpool_sda
> ignore_lv_errors=no
> vgname=gluster_vg_sda
> lvtype=thinpool
> size=1005GB
> poolmetadatasize=5GB
>
> [lv2:cmdnode2.cmd911.com]
> action=create
> poolname=gluster_thinpool_sdb
> ignore_lv_errors=no
> vgname=gluster_vg_sdb
> lvtype=thinpool
> size=1005GB
> poolmetadatasize=5GB
>
> [lv3:cmdnode3.cmd911.com]
> action=create
> poolname=gluster_thinpool_sdb
> ignore_lv_errors=no
> vgname=gluster_vg_sdb
> lvtype=thinpool
> size=41GB
> poolmetadatasize=1GB
>
> [lv4:cmdnode1.cmd911.com]
> action=create
> lvname=gluster_lv_engine
> ignore_lv_errors=no
> vgname=gluster_vg_sda
> mount=/gluster_bricks/engine
> size=100GB
> lvtype=thick
>
> [lv5:cmdnode1.cmd911.com]
> action=create
> lvname=gluster_lv_data
> ignore_lv_errors=no
> vgname=gluster_vg_sda
> mount=/gluster_bricks/data
> lvtype=thinlv
> poolname=gluster_thinpool_sda
> virtualsize=500GB
>
> [lv6:cmdnode1.cmd911.com]
> action=create
> lvname=gluster_lv_vmstore
> ignore_lv_errors=no
> vgname=gluster_vg_sda
> mount=/gluster_bricks/vmstore
> lvtype=thinlv
> poolname=gluster_thinpool_sda
> virtualsize=500GB
>
> [lv7:cmdnode2.cmd911.com]
> action=create
> lvname=gluster_lv_engine
> ignore_lv_errors=no
> vgname=gluster_vg_sdb
> mount=/gluster_bricks/engine
> size=100GB
> lvtype=thick
>
> [lv8:cmdnode2.cmd911.com]
> action=create
> lvname=gluster_lv_data
> ignore_lv_errors=no
> vgname=gluster_vg_sdb
> mount=/gluster_bricks/data
> lvtype=thinlv
> poolname=gluster_thinpool_sdb
> virtualsize=500GB
>
> [lv9:cmdnode2.cmd911.com]
> action=create
> lvname=gluster_lv_vmstore
> ignore_lv_errors=no
> vgname=gluster_vg_sdb
> mount=/gluster_bricks/vmstore
> lvtype=thinlv
> poolname=gluster_thinpool_sdb
> virtualsize=500GB
>
> [lv10:cmdnode3.cmd911.com]
> action=create
> lvname=gluster_lv_engine
> ignore_lv_errors=no
> vgname=gluster_vg_sdb
> mount=/gluster_bricks/engine
> size=20GB
> lvtype=thick
>
> [lv11:cmdnode3.cmd911.com]
> action=create
> lvname=gluster_lv_data
> ignore_lv_errors=no
> vgname=gluster_vg_sdb
> mount=/gluster_bricks/data
> lvtype=thinlv
> poolname=gluster_thinpool_sdb
> virtualsize=20GB
>
> [lv12:cmdnode3.cmd911.com]
> action=create
> lvname=gluster_lv_vmstore
> ignore_lv_errors=no
> vgname=gluster_vg_sdb
> mount=/gluster_bricks/vmstore
> lvtype=thinlv
> poolname=gluster_thinpool_sdb
> virtualsize=20GB
>
> [selinux]
> yes
>
> [service3]
> action=restart
> service=glusterd
> slice_setup=yes
>
> [firewalld]
> action=add
> ports=111/tcp,2049/tcp,54321/tcp,5900/tcp,5900-6923/tcp,5666/tcp,16514/tcp,54322/tcp
> services=glusterfs
>
> [script2]
> action=execute
> file=/usr/share/gdeploy/scripts/disable-gluster-hooks.sh
>
> [shell3]
> action=execute
> command=usermod -a -G gluster qemu
>
> [volume1]
> action=create
> volname=engine
> transport=tcp
> replica=yes
> replica_count=3
> key=group,storage.owner-uid,storage.owner-gid,network.ping-timeout,performance.strict-o-direct,network.remote-dio,cluster.granular-entry-heal
> value=virt,36,36,30,on,off,enable
> brick_dirs=cmdnode1.cmd911.com:/gluster_bricks/engine/engine,cmdnode2.cmd911.com:/gluster_bricks/engine/engine,cmdnode3.cmd911.com:/gluster_bricks/engine/engine
> ignore_volume_errors=no
> arbiter_count=1
>
> [volume2]
> action=create
> volname=data
> transport=tcp
> replica=yes
> replica_count=3
> key=group,storage.owner-uid,storage.owner-gid,network.ping-timeout,performance.strict-o-direct,network.remote-dio,cluster.granular-entry-heal
> value=virt,36,36,30,on,off,enable
> brick_dirs=cmdnode1.cmd911.com:/gluster_bricks/data/data,cmdnode2.cmd911.com:/gluster_bricks/data/data,cmdnode3.cmd911.com:/gluster_bricks/data/data
> ignore_volume_errors=no
> arbiter_count=1
>
> [volume3]
> action=create
> volname=vmstore
> transport=tcp
> replica=yes
> replica_count=3
> key=group,storage.owner-uid,storage.owner-gid,network.ping-timeout,performance.strict-o-direct,network.remote-dio,cluster.granular-entry-heal
> value=virt,36,36,30,on,off,enable
> brick_dirs=cmdnode1.cmd911.com:/gluster_bricks/vmstore/vmstore,cmdnode2.cmd911.com:/gluster_bricks/vmstore/vmstore,cmdnode3.cmd911.com:/gluster_bricks/vmstore/vmstore
> ignore_volume_errors=no
> arbiter_count=1
> _______________________________________________
> Users mailing list -- users@ovirt.org
> To unsubscribe send an email to users-leave@ovirt.org
> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/
> List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/5UA6OAQ2I7QJPCM6KQTLXG2NXVTYOTHD/