
I got the Gluster working but I can’t finish setting up the engine it fails and says can’t query dns for engine and then can’t remove the storage domain without redoing everything all over again. On Tue, Feb 26, 2019 at 2:27 AM Parth Dhanjal <dparth@redhat.com> wrote:
Hey Matthew!
Can you please provide me with the following to help you debug the issue that you are facing? 1. oVirt and gdeploy version 2. /var/log/messages file 3. /root/.gdeploy file
On Mon, Feb 25, 2019 at 1:23 PM Parth Dhanjal <dparth@redhat.com> wrote:
Hey Matthew!
Can you please provide which oVirt and gdeploy version have you installed?
Regards Parth Dhanjal
On Mon, Feb 25, 2019 at 12:56 PM Sahina Bose <sabose@redhat.com> wrote:
+Gobinda Das +Dhanjal Parth can you please check?
On Fri, Feb 22, 2019 at 11:52 PM Matthew Roth <mroth381@gmail.com> wrote:
I have 3 servers, Node 1 is 3tb /dev/sda, Node 2, 3tb /dev/sdb,
node3 3tb /dev/sdb
I start the process for gluster deployment. I change node 1 to sda and
all the other ones to sdb. I get no errors however,
when I get to Creating physical Volume all it does is spin forever . doesnt get any
further. I can leave it there for 5 hours and doesn't go anywhere.
#gdeploy configuration generated by cockpit-gluster plugin [hosts] cmdnode1.cmd911.com cmdnode2.cmd911.com cmdnode3.cmd911.com
[script1:cmdnode1.cmd911.com] action=execute ignore_script_errors=no file=/usr/share/gdeploy/scripts/grafton-sanity-check.sh -d sda -h
cmdnode1.cmd911.com, cmdnode2.cmd911.com, cmdnode3.cmd911.com
[script1:cmdnode2.cmd911.com] action=execute ignore_script_errors=no file=/usr/share/gdeploy/scripts/grafton-sanity-check.sh -d sdb -h
cmdnode1.cmd911.com, cmdnode2.cmd911.com, cmdnode3.cmd911.com
[script1:cmdnode3.cmd911.com] action=execute ignore_script_errors=no file=/usr/share/gdeploy/scripts/grafton-sanity-check.sh -d sdb -h
cmdnode1.cmd911.com, cmdnode2.cmd911.com, cmdnode3.cmd911.com
[disktype] raid6
[diskcount] 12
[stripesize] 256
[service1] action=enable service=chronyd
[service2] action=restart service=chronyd
[shell2] action=execute command=vdsm-tool configure --force
[script3] action=execute file=/usr/share/gdeploy/scripts/blacklist_all_disks.sh ignore_script_errors=no
[pv1:cmdnode1.cmd911.com] action=create devices=sda ignore_pv_errors=no
[pv1:cmdnode2.cmd911.com] action=create devices=sdb ignore_pv_errors=no
[pv1:cmdnode3.cmd911.com] action=create devices=sdb ignore_pv_errors=no
[vg1:cmdnode1.cmd911.com] action=create vgname=gluster_vg_sda pvname=sda ignore_vg_errors=no
[vg1:cmdnode2.cmd911.com] action=create vgname=gluster_vg_sdb pvname=sdb ignore_vg_errors=no
[vg1:cmdnode3.cmd911.com] action=create vgname=gluster_vg_sdb pvname=sdb ignore_vg_errors=no
[lv1:cmdnode1.cmd911.com] action=create poolname=gluster_thinpool_sda ignore_lv_errors=no vgname=gluster_vg_sda lvtype=thinpool size=1005GB poolmetadatasize=5GB
[lv2:cmdnode2.cmd911.com] action=create poolname=gluster_thinpool_sdb ignore_lv_errors=no vgname=gluster_vg_sdb lvtype=thinpool size=1005GB poolmetadatasize=5GB
[lv3:cmdnode3.cmd911.com] action=create poolname=gluster_thinpool_sdb ignore_lv_errors=no vgname=gluster_vg_sdb lvtype=thinpool size=41GB poolmetadatasize=1GB
[lv4:cmdnode1.cmd911.com] action=create lvname=gluster_lv_engine ignore_lv_errors=no vgname=gluster_vg_sda mount=/gluster_bricks/engine size=100GB lvtype=thick
[lv5:cmdnode1.cmd911.com] action=create lvname=gluster_lv_data ignore_lv_errors=no vgname=gluster_vg_sda mount=/gluster_bricks/data lvtype=thinlv poolname=gluster_thinpool_sda virtualsize=500GB
[lv6:cmdnode1.cmd911.com] action=create lvname=gluster_lv_vmstore ignore_lv_errors=no vgname=gluster_vg_sda mount=/gluster_bricks/vmstore lvtype=thinlv poolname=gluster_thinpool_sda virtualsize=500GB
[lv7:cmdnode2.cmd911.com] action=create lvname=gluster_lv_engine ignore_lv_errors=no vgname=gluster_vg_sdb mount=/gluster_bricks/engine size=100GB lvtype=thick
[lv8:cmdnode2.cmd911.com] action=create lvname=gluster_lv_data ignore_lv_errors=no vgname=gluster_vg_sdb mount=/gluster_bricks/data lvtype=thinlv poolname=gluster_thinpool_sdb virtualsize=500GB
[lv9:cmdnode2.cmd911.com] action=create lvname=gluster_lv_vmstore ignore_lv_errors=no vgname=gluster_vg_sdb mount=/gluster_bricks/vmstore lvtype=thinlv poolname=gluster_thinpool_sdb virtualsize=500GB
[lv10:cmdnode3.cmd911.com] action=create lvname=gluster_lv_engine ignore_lv_errors=no vgname=gluster_vg_sdb mount=/gluster_bricks/engine size=20GB lvtype=thick
[lv11:cmdnode3.cmd911.com] action=create lvname=gluster_lv_data ignore_lv_errors=no vgname=gluster_vg_sdb mount=/gluster_bricks/data lvtype=thinlv poolname=gluster_thinpool_sdb virtualsize=20GB
[lv12:cmdnode3.cmd911.com] action=create lvname=gluster_lv_vmstore ignore_lv_errors=no vgname=gluster_vg_sdb mount=/gluster_bricks/vmstore lvtype=thinlv poolname=gluster_thinpool_sdb virtualsize=20GB
[selinux] yes
[service3] action=restart service=glusterd slice_setup=yes
[firewalld] action=add
ports=111/tcp,2049/tcp,54321/tcp,5900/tcp,5900-6923/tcp,5666/tcp,16514/tcp,54322/tcp
services=glusterfs
[script2] action=execute file=/usr/share/gdeploy/scripts/disable-gluster-hooks.sh
[shell3] action=execute command=usermod -a -G gluster qemu
[volume1] action=create volname=engine transport=tcp replica=yes replica_count=3
key=group,storage.owner-uid,storage.owner-gid,network.ping-timeout,performance.strict-o-direct,network.remote-dio,cluster.granular-entry-heal
value=virt,36,36,30,on,off,enable brick_dirs=cmdnode1.cmd911.com: /gluster_bricks/engine/engine,cmdnode2.cmd911.com: /gluster_bricks/engine/engine,cmdnode3.cmd911.com: /gluster_bricks/engine/engine ignore_volume_errors=no arbiter_count=1
[volume2] action=create volname=data transport=tcp replica=yes replica_count=3
key=group,storage.owner-uid,storage.owner-gid,network.ping-timeout,performance.strict-o-direct,network.remote-dio,cluster.granular-entry-heal
value=virt,36,36,30,on,off,enable brick_dirs=cmdnode1.cmd911.com: /gluster_bricks/data/data,cmdnode2.cmd911.com: /gluster_bricks/data/data,cmdnode3.cmd911.com:/gluster_bricks/data/data ignore_volume_errors=no arbiter_count=1
[volume3] action=create volname=vmstore transport=tcp replica=yes replica_count=3
key=group,storage.owner-uid,storage.owner-gid,network.ping-timeout,performance.strict-o-direct,network.remote-dio,cluster.granular-entry-heal
value=virt,36,36,30,on,off,enable brick_dirs=cmdnode1.cmd911.com: /gluster_bricks/vmstore/vmstore,cmdnode2.cmd911.com: /gluster_bricks/vmstore/vmstore,cmdnode3.cmd911.com: /gluster_bricks/vmstore/vmstore ignore_volume_errors=no arbiter_count=1 _______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-leave@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/5UA6OAQ2I7QJPC...