I'm trying to bring up a single node hyperconverged with the current
node-ng ISO installation,
but it ends with this failure message.
TASK [gluster.features/roles/gluster_hci : Check if /var/log has enough
disk space] ***
fatal: [
br014.bridges.psc.edu]: FAILED! => {"changed": true, "cmd":
"df -m
/var/log | awk '/[0-9]%/ {print $4}'", "delta":
"0:00:00.008513", "end":
"2019-05-09 20:09:27.914400", "failed_when_result": true,
"rc": 0, "start":
"2019-05-09 20:09:27.905887", "stderr": "",
"stderr_lines": [], "stdout":
"7470", "stdout_lines": ["7470"]}
I have what the installer created by default for /var/log, so I don't know
why its complaining.
[root@br014 ~]# df -kh
Filesystem Size Used
Avail Use% Mounted on
/dev/mapper/onn_br014-ovirt--node--ng--4.3.3.1--0.20190417.0+1 3.5T 2.1G
3.3T 1% /
devtmpfs 63G
0 63G 0% /dev
tmpfs 63G
4.0K 63G 1% /dev/shm
tmpfs 63G
18M 63G 1% /run
tmpfs 63G
0 63G 0% /sys/fs/cgroup
/dev/mapper/onn_br014-home 976M 2.6M
907M 1% /home
/dev/mapper/onn_br014-tmp 976M 2.8M
906M 1% /tmp
/dev/mapper/onn_br014-var 15G
42M 14G 1% /var
/dev/sda2 976M 173M
737M 19% /boot
/dev/mapper/onn_br014-var_log 7.8G 41M
7.3G 1% /var/log
/dev/mapper/onn_br014-var_log_audit 2.0G 7.6M
1.8G 1% /var/log/audit
/dev/mapper/onn_br014-var_crash 9.8G 37M
9.2G 1% /var/crash
/dev/sda1 200M 12M
189M 6% /boot/efi
tmpfs 13G
0 13G 0% /run/user/1000
tmpfs 13G
0 13G 0% /run/user/0
/dev/mapper/gluster_vg_sdb-gluster_lv_engine 3.7T 33M
3.7T 1% /gluster_bricks/engine
/dev/mapper/gluster_vg_sdc-gluster_lv_data 3.7T 34M
3.7T 1% /gluster_bricks/data
/dev/mapper/gluster_vg_sdd-gluster_lv_vmstore 3.7T 34M
3.7T 1% /gluster_bricks/vmstore
The machine had 4 4TB disks, so sda is the installation for oVirt node-ng,
the other 3 disks for the gluster volumes.
root@br014 ~]# pvs
PV VG Fmt Attr PSize PFree
/dev/sda3 onn_br014 lvm2 a-- <3.64t 100.00g
/dev/sdb gluster_vg_sdb lvm2 a-- <3.64t <26.02g
/dev/sdc gluster_vg_sdc lvm2 a-- <3.64t 0
/dev/sdd gluster_vg_sdd lvm2 a-- <3.64t 0
[root@br014 ~]# vgs
VG #PV #LV #SN Attr VSize VFree
gluster_vg_sdb 1 1 0 wz--n- <3.64t <26.02g
gluster_vg_sdc 1 2 0 wz--n- <3.64t 0
gluster_vg_sdd 1 2 0 wz--n- <3.64t 0
onn_br014 1 11 0 wz--n- <3.64t 100.00g
[root@br014 ~]# lvs
LV VG Attr LSize
Pool Origin Data%
Meta% Move Log Cpy%Sync Convert
gluster_lv_engine gluster_vg_sdb -wi-ao----
3.61t
gluster_lv_data gluster_vg_sdc Vwi-aot--- 3.61t
gluster_thinpool_gluster_vg_sdc
0.05
gluster_thinpool_gluster_vg_sdc gluster_vg_sdc twi-aot---
<3.61t
0.05 0.13
gluster_lv_vmstore gluster_vg_sdd Vwi-aot--- 3.61t
gluster_thinpool_gluster_vg_sdd
0.05
gluster_thinpool_gluster_vg_sdd gluster_vg_sdd twi-aot---
<3.61t
0.05 0.13
home onn_br014 Vwi-aotz-- 1.00g
pool00
4.79
ovirt-node-ng-4.3.3.1-0.20190417.0 onn_br014 Vwi---tz-k <3.51t
pool00
root
ovirt-node-ng-4.3.3.1-0.20190417.0+1 onn_br014 Vwi-aotz-- <3.51t
pool00 ovirt-node-ng-4.3.3.1-0.20190417.0
0.13
pool00 onn_br014 twi-aotz--
3.53t
0.19 1.86
root onn_br014 Vri---tz-k <3.51t
pool00
swap onn_br014 -wi-ao----
4.00g
tmp onn_br014 Vwi-aotz-- 1.00g
pool00
4.84
var onn_br014 Vwi-aotz-- 15.00g
pool00
3.67
var_crash onn_br014 Vwi-aotz-- 10.00g
pool00
2.86
var_log onn_br014 Vwi-aotz-- 8.00g
pool00
3.25
var_log_audit onn_br014 Vwi-aotz-- 2.00g
pool00
4.86
Here's the full deploy log from the UI. Let me know if you need specific
logs.
PLAY [Setup backend]
***********************************************************
TASK [Gathering Facts]
*********************************************************
ok: [
br014.bridges.psc.edu]
TASK [gluster.infra/roles/firewall_config : Start firewalld if not already
started] ***
ok: [
br014.bridges.psc.edu]
TASK [gluster.infra/roles/firewall_config : check if required variables are
set] ***
skipping: [
br014.bridges.psc.edu]
TASK [gluster.infra/roles/firewall_config : Open/Close firewalld ports]
********
ok: [
br014.bridges.psc.edu] => (item=2049/tcp)
ok: [
br014.bridges.psc.edu] => (item=54321/tcp)
ok: [
br014.bridges.psc.edu] => (item=5900/tcp)
ok: [
br014.bridges.psc.edu] => (item=5900-6923/tcp)
ok: [
br014.bridges.psc.edu] => (item=5666/tcp)
ok: [
br014.bridges.psc.edu] => (item=16514/tcp)
TASK [gluster.infra/roles/firewall_config : Add/Delete services to
firewalld rules] ***
ok: [
br014.bridges.psc.edu] => (item=glusterfs)
TASK [gluster.infra/roles/backend_setup : Gather facts to determine the OS
distribution] ***
ok: [
br014.bridges.psc.edu]
TASK [gluster.infra/roles/backend_setup : Change to Install lvm tools for
debian systems.] ***
skipping: [
br014.bridges.psc.edu]
TASK [gluster.infra/roles/backend_setup : Change to Install lvm tools for
RHEL systems.] ***
ok: [
br014.bridges.psc.edu]
TASK [gluster.infra/roles/backend_setup : Install python-yaml package for
Debian systems] ***
skipping: [
br014.bridges.psc.edu]
TASK [gluster.infra/roles/backend_setup : Initialize vdo_devs array]
***********
ok: [
br014.bridges.psc.edu]
TASK [gluster.infra/roles/backend_setup : Record VDO devices (if any)]
*********
skipping: [
br014.bridges.psc.edu] => (item={u'vgname':
u'gluster_vg_sdb',
u'pvname': u'/dev/sdb'})
skipping: [
br014.bridges.psc.edu] => (item={u'vgname':
u'gluster_vg_sdc',
u'pvname': u'/dev/sdc'})
skipping: [
br014.bridges.psc.edu] => (item={u'vgname':
u'gluster_vg_sdd',
u'pvname': u'/dev/sdd'})
TASK [gluster.infra/roles/backend_setup : Enable and start vdo service]
********
skipping: [
br014.bridges.psc.edu]
TASK [gluster.infra/roles/backend_setup : Create VDO with specified size]
******
skipping: [
br014.bridges.psc.edu]
TASK [gluster.infra/roles/backend_setup : Check if valid disktype is
provided] ***
skipping: [
br014.bridges.psc.edu]
TASK [gluster.infra/roles/backend_setup : Set PV data alignment for JBOD]
******
ok: [
br014.bridges.psc.edu]
TASK [gluster.infra/roles/backend_setup : Set PV data alignment for RAID]
******
skipping: [
br014.bridges.psc.edu]
TASK [gluster.infra/roles/backend_setup : Set VG physical extent size for
RAID] ***
skipping: [
br014.bridges.psc.edu]
TASK [gluster.infra/roles/backend_setup : Create volume groups]
****************
ok: [
br014.bridges.psc.edu] => (item={u'vgname': u'gluster_vg_sdb',
u'pvname': u'/dev/sdb'})
ok: [
br014.bridges.psc.edu] => (item={u'vgname': u'gluster_vg_sdc',
u'pvname': u'/dev/sdc'})
ok: [
br014.bridges.psc.edu] => (item={u'vgname': u'gluster_vg_sdd',
u'pvname': u'/dev/sdd'})
TASK [gluster.infra/roles/backend_setup : Create thick logical volume]
*********
ok: [
br014.bridges.psc.edu] => (item={u'lvname': u'gluster_lv_engine',
u'vgname': u'gluster_vg_sdb', u'size': u'3700G'})
TASK [gluster.infra/roles/backend_setup : Calculate chunksize for
RAID6/RAID10/RAID5] ***
skipping: [
br014.bridges.psc.edu]
TASK [gluster.infra/roles/backend_setup : Set chunksize for JBOD]
**************
ok: [
br014.bridges.psc.edu]
TASK [gluster.infra/roles/backend_setup : Create a LV thinpool]
****************
ok: [
br014.bridges.psc.edu] => (item={u'vgname': u'gluster_vg_sdc',
u'thinpoolname': u'gluster_thinpool_gluster_vg_sdc',
u'poolmetadatasize':
u'16G'})
ok: [
br014.bridges.psc.edu] => (item={u'vgname': u'gluster_vg_sdd',
u'thinpoolname': u'gluster_thinpool_gluster_vg_sdd',
u'poolmetadatasize':
u'16G'})
TASK [gluster.infra/roles/backend_setup : Create thin logical volume]
**********
ok: [
br014.bridges.psc.edu] => (item={u'lvname': u'gluster_lv_data',
u'vgname': u'gluster_vg_sdc', u'thinpool':
u'gluster_thinpool_gluster_vg_sdc', u'lvsize': u'3700G'})
ok: [
br014.bridges.psc.edu] => (item={u'lvname':
u'gluster_lv_vmstore',
u'vgname': u'gluster_vg_sdd', u'thinpool':
u'gluster_thinpool_gluster_vg_sdd', u'lvsize': u'3700G'})
TASK [gluster.infra/roles/backend_setup : Extend volume group]
*****************
skipping: [
br014.bridges.psc.edu]
TASK [gluster.infra/roles/backend_setup : Change attributes of LV]
*************
skipping: [
br014.bridges.psc.edu]
TASK [gluster.infra/roles/backend_setup : Create LV for cache]
*****************
skipping: [
br014.bridges.psc.edu]
TASK [gluster.infra/roles/backend_setup : Create metadata LV for cache]
********
skipping: [
br014.bridges.psc.edu]
TASK [gluster.infra/roles/backend_setup : Convert logical volume to a cache
pool LV] ***
skipping: [
br014.bridges.psc.edu]
TASK [gluster.infra/roles/backend_setup : Convert logical volume to a cache
pool LV without cachemetalvname] ***
skipping: [
br014.bridges.psc.edu]
TASK [gluster.infra/roles/backend_setup : Convert an existing logical
volume to a cache LV] ***
skipping: [
br014.bridges.psc.edu]
TASK [gluster.infra/roles/backend_setup : Set XFS options for JBOD]
************
ok: [
br014.bridges.psc.edu]
TASK [gluster.infra/roles/backend_setup : Set XFS options for RAID devices]
****
skipping: [
br014.bridges.psc.edu]
TASK [gluster.infra/roles/backend_setup : Create filesystem on thin logical
vols] ***
ok: [
br014.bridges.psc.edu] => (item={u'lvname': u'gluster_lv_data',
u'vgname': u'gluster_vg_sdc', u'thinpool':
u'gluster_thinpool_gluster_vg_sdc', u'lvsize': u'3700G'})
ok: [
br014.bridges.psc.edu] => (item={u'lvname':
u'gluster_lv_vmstore',
u'vgname': u'gluster_vg_sdd', u'thinpool':
u'gluster_thinpool_gluster_vg_sdd', u'lvsize': u'3700G'})
TASK [gluster.infra/roles/backend_setup : Create filesystem on thick
logical vols] ***
ok: [
br014.bridges.psc.edu] => (item={u'lvname': u'gluster_lv_engine',
u'vgname': u'gluster_vg_sdb', u'size': u'3700G'})
TASK [gluster.infra/roles/backend_setup : Create mount directories if not
already present] ***
ok: [
br014.bridges.psc.edu] => (item={u'path':
u'/gluster_bricks/engine',
u'vgname': u'gluster_vg_sdb', u'lvname':
u'gluster_lv_engine'})
ok: [
br014.bridges.psc.edu] => (item={u'path':
u'/gluster_bricks/data',
u'vgname': u'gluster_vg_sdc', u'lvname':
u'gluster_lv_data'})
ok: [
br014.bridges.psc.edu] => (item={u'path':
u'/gluster_bricks/vmstore',
u'vgname': u'gluster_vg_sdd', u'lvname':
u'gluster_lv_vmstore'})
TASK [gluster.infra/roles/backend_setup : Set mount options for VDO]
***********
skipping: [
br014.bridges.psc.edu]
TASK [gluster.infra/roles/backend_setup : Mount the vdo devices (If any)]
******
skipping: [
br014.bridges.psc.edu] => (item={u'path':
u'/gluster_bricks/engine', u'vgname': u'gluster_vg_sdb',
u'lvname':
u'gluster_lv_engine'})
skipping: [
br014.bridges.psc.edu] => (item={u'path':
u'/gluster_bricks/data', u'vgname': u'gluster_vg_sdc',
u'lvname':
u'gluster_lv_data'})
skipping: [
br014.bridges.psc.edu] => (item={u'path':
u'/gluster_bricks/vmstore', u'vgname': u'gluster_vg_sdd',
u'lvname':
u'gluster_lv_vmstore'})
TASK [gluster.infra/roles/backend_setup : Mount the devices]
*******************
ok: [
br014.bridges.psc.edu] => (item={u'path':
u'/gluster_bricks/engine',
u'vgname': u'gluster_vg_sdb', u'lvname':
u'gluster_lv_engine'})
ok: [
br014.bridges.psc.edu] => (item={u'path':
u'/gluster_bricks/data',
u'vgname': u'gluster_vg_sdc', u'lvname':
u'gluster_lv_data'})
ok: [
br014.bridges.psc.edu] => (item={u'path':
u'/gluster_bricks/vmstore',
u'vgname': u'gluster_vg_sdd', u'lvname':
u'gluster_lv_vmstore'})
TASK [gluster.infra/roles/backend_setup : Set Gluster specific SeLinux
context on the bricks] ***
ok: [
br014.bridges.psc.edu] => (item={u'path':
u'/gluster_bricks/engine',
u'vgname': u'gluster_vg_sdb', u'lvname':
u'gluster_lv_engine'})
ok: [
br014.bridges.psc.edu] => (item={u'path':
u'/gluster_bricks/data',
u'vgname': u'gluster_vg_sdc', u'lvname':
u'gluster_lv_data'})
ok: [
br014.bridges.psc.edu] => (item={u'path':
u'/gluster_bricks/vmstore',
u'vgname': u'gluster_vg_sdd', u'lvname':
u'gluster_lv_vmstore'})
TASK [gluster.infra/roles/backend_setup : restore file(s) default SELinux
security contexts] ***
changed: [
br014.bridges.psc.edu] => (item={u'path':
u'/gluster_bricks/engine', u'vgname': u'gluster_vg_sdb',
u'lvname':
u'gluster_lv_engine'})
changed: [
br014.bridges.psc.edu] => (item={u'path':
u'/gluster_bricks/data', u'vgname': u'gluster_vg_sdc',
u'lvname':
u'gluster_lv_data'})
changed: [
br014.bridges.psc.edu] => (item={u'path':
u'/gluster_bricks/vmstore', u'vgname': u'gluster_vg_sdd',
u'lvname':
u'gluster_lv_vmstore'})
TASK [gluster.infra/roles/backend_reset : unmount the directories (if
mounted)] ***
skipping: [
br014.bridges.psc.edu]
TASK [gluster.infra/roles/backend_reset : Delete volume groups]
****************
skipping: [
br014.bridges.psc.edu]
TASK [gluster.infra/roles/backend_reset : Remove VDO devices]
******************
skipping: [
br014.bridges.psc.edu]
TASK [gluster.features/roles/nfs_ganesha : Create temporary storage
directory] ***
skipping: [
br014.bridges.psc.edu]
TASK [gluster.features/roles/nfs_ganesha : Get the name of the directory
created] ***
skipping: [
br014.bridges.psc.edu]
TASK [gluster.features/roles/nfs_ganesha : check if
gluster_features_ganesha_clusternodes is set] ***
skipping: [
br014.bridges.psc.edu]
TASK [gluster.features/roles/nfs_ganesha : Define service port]
****************
skipping: [
br014.bridges.psc.edu] => (item=^#(STATD_PORT=.*))
skipping: [
br014.bridges.psc.edu] => (item=^#(LOCKD_TCPPORT=.*))
skipping: [
br014.bridges.psc.edu] => (item=^#(LOCKD_UDPPORT=.*))
TASK [gluster.features/roles/nfs_ganesha : Check packages installed, if not
install] ***
skipping: [
br014.bridges.psc.edu] => (item=glusterfs-ganesha)
skipping: [
br014.bridges.psc.edu] => (item=nfs-ganesha)
skipping: [
br014.bridges.psc.edu] => (item=corosync)
skipping: [
br014.bridges.psc.edu] => (item=pacemaker)
skipping: [
br014.bridges.psc.edu] => (item=libntirpc)
skipping: [
br014.bridges.psc.edu] => (item=pcs)
TASK [gluster.features/roles/nfs_ganesha : Restart services]
*******************
skipping: [
br014.bridges.psc.edu] => (item=nfslock)
skipping: [
br014.bridges.psc.edu] => (item=nfs-config)
skipping: [
br014.bridges.psc.edu] => (item=rpc-statd)
TASK [gluster.features/roles/nfs_ganesha : Stop services]
**********************
skipping: [
br014.bridges.psc.edu] => (item=nfs-server)
TASK [gluster.features/roles/nfs_ganesha : Disable service]
********************
skipping: [
br014.bridges.psc.edu] => (item=nfs-server)
TASK [gluster.features/roles/nfs_ganesha : Enable services]
********************
skipping: [
br014.bridges.psc.edu] => (item=glusterfssharedstorage)
skipping: [
br014.bridges.psc.edu] => (item=nfs-ganesha)
skipping: [
br014.bridges.psc.edu] => (item=network)
skipping: [
br014.bridges.psc.edu] => (item=pcsd)
skipping: [
br014.bridges.psc.edu] => (item=pacemaker)
TASK [gluster.features/roles/nfs_ganesha : Start services]
*********************
skipping: [
br014.bridges.psc.edu] => (item=network)
skipping: [
br014.bridges.psc.edu] => (item=pcsd)
TASK [gluster.features/roles/nfs_ganesha : Create a user hacluster if not
already present] ***
skipping: [
br014.bridges.psc.edu]
TASK [gluster.features/roles/nfs_ganesha : Set the password for hacluster]
*****
skipping: [
br014.bridges.psc.edu]
TASK [gluster.features/roles/nfs_ganesha : Set the hacluster user the same
password on new nodes] ***
skipping: [
br014.bridges.psc.edu]
TASK [gluster.features/roles/nfs_ganesha : Pcs cluster authenticate the
hacluster on new nodes] ***
skipping: [
br014.bridges.psc.edu]
TASK [gluster.features/roles/nfs_ganesha : Pause for a few seconds after
pcs auth] ***
skipping: [
br014.bridges.psc.edu]
TASK [gluster.features/roles/nfs_ganesha : Set gluster_use_execmem flag on
and keep it persistent] ***
skipping: [
br014.bridges.psc.edu] => (item=gluster_use_execmem)
skipping: [
br014.bridges.psc.edu] => (item=ganesha_use_fusefs)
TASK [gluster.features/roles/nfs_ganesha : check if
gluster_features_ganesha_masternode is set] ***
skipping: [
br014.bridges.psc.edu]
TASK [gluster.features/roles/nfs_ganesha : Copy the ssh keys to the local
machine] ***
skipping: [
br014.bridges.psc.edu] => (item=secret.pem.pub)
skipping: [
br014.bridges.psc.edu] => (item=secret.pem)
TASK [gluster.features/roles/nfs_ganesha : check if
gluster_features_ganesha_newnodes_vip is set] ***
skipping: [
br014.bridges.psc.edu]
TASK [gluster.features/roles/nfs_ganesha : Copy the public key to remote
nodes] ***
skipping: [
br014.bridges.psc.edu]
TASK [gluster.features/roles/nfs_ganesha : Copy the private key to remote
node] ***
skipping: [
br014.bridges.psc.edu]
TASK [gluster.features/roles/nfs_ganesha : Deploy the pubkey on all nodes]
*****
skipping: [
br014.bridges.psc.edu]
TASK [gluster.features/roles/nfs_ganesha : Make the volume a gluster shared
volume] ***
skipping: [
br014.bridges.psc.edu]
TASK [gluster.features/roles/nfs_ganesha : Generate ssh key in one of the
nodes in HA cluster] ***
skipping: [
br014.bridges.psc.edu]
TASK [gluster.features/roles/nfs_ganesha : Copy the ssh keys to the local
machine] ***
skipping: [
br014.bridges.psc.edu] => (item=secret.pem.pub)
skipping: [
br014.bridges.psc.edu] => (item=secret.pem)
TASK [gluster.features/roles/nfs_ganesha : Create configuration directory
for nfs_ganesha] ***
skipping: [
br014.bridges.psc.edu]
TASK [gluster.features/roles/nfs_ganesha : Copy ganesha.conf to config
directory on shared volume] ***
skipping: [
br014.bridges.psc.edu]
TASK [gluster.features/roles/nfs_ganesha : Create ganesha-ha.conf file]
********
skipping: [
br014.bridges.psc.edu]
TASK [gluster.features/roles/nfs_ganesha : Enable NFS Ganesha]
*****************
skipping: [
br014.bridges.psc.edu]
TASK [gluster.features/roles/nfs_ganesha : Pause for 30 seconds (takes a
while to enable NFS Ganesha)] ***
skipping: [
br014.bridges.psc.edu]
TASK [gluster.features/roles/nfs_ganesha : Check NFS Ganesha status]
***********
skipping: [
br014.bridges.psc.edu]
TASK [gluster.features/roles/nfs_ganesha : Report NFS Ganesha status]
**********
skipping: [
br014.bridges.psc.edu]
TASK [gluster.features/roles/nfs_ganesha : Report NFS Ganesha status (If
any errors)] ***
skipping: [
br014.bridges.psc.edu]
TASK [gluster.features/roles/nfs_ganesha : check if
gluster_features_ganesha_volume is set] ***
skipping: [
br014.bridges.psc.edu]
TASK [gluster.features/roles/nfs_ganesha : Export the NFS Ganesha volume]
******
skipping: [
br014.bridges.psc.edu]
TASK [gluster.features/roles/nfs_ganesha : Copy the public key to remote
nodes] ***
skipping: [
br014.bridges.psc.edu]
TASK [gluster.features/roles/nfs_ganesha : Copy the private key to remote
node] ***
skipping: [
br014.bridges.psc.edu]
TASK [gluster.features/roles/nfs_ganesha : Deploy the pubkey on all nodes]
*****
skipping: [
br014.bridges.psc.edu]
TASK [gluster.features/roles/nfs_ganesha : Adds a node to the cluster]
*********
skipping: [
br014.bridges.psc.edu]
TASK [gluster.features/roles/nfs_ganesha : Report ganesha add-node status]
*****
skipping: [
br014.bridges.psc.edu]
TASK [gluster.features/roles/nfs_ganesha : Delete the temporary directory]
*****
skipping: [
br014.bridges.psc.edu]
TASK [gluster.features/roles/gluster_hci : Check if packages are installed,
if not install] ***
ok: [
br014.bridges.psc.edu] => (item=vdsm)
ok: [
br014.bridges.psc.edu] => (item=vdsm-gluster)
ok: [
br014.bridges.psc.edu] => (item=ovirt-host)
ok: [
br014.bridges.psc.edu] => (item=screen)
TASK [gluster.features/roles/gluster_hci : Enable and start glusterd and
chronyd] ***
ok: [
br014.bridges.psc.edu] => (item=chronyd)
ok: [
br014.bridges.psc.edu] => (item=glusterd)
ok: [
br014.bridges.psc.edu] => (item=firewalld)
TASK [gluster.features/roles/gluster_hci : Add user qemu to gluster group]
*****
ok: [
br014.bridges.psc.edu]
TASK [gluster.features/roles/gluster_hci : Disable the hook scripts]
***********
changed: [
br014.bridges.psc.edu] =>
(item=/var/lib/glusterd/hooks/1/set/post/S30samba-set.sh)
changed: [
br014.bridges.psc.edu] =>
(item=/var/lib/glusterd/hooks/1/start/post/S30samba-start.sh)
changed: [
br014.bridges.psc.edu] =>
(item=/var/lib/glusterd/hooks/1/stop/pre/S30samba-stop.sh)
changed: [
br014.bridges.psc.edu] =>
(item=/var/lib/glusterd/hooks/1/reset/post/S31ganesha-reset.sh)
changed: [
br014.bridges.psc.edu] =>
(item=/var/lib/glusterd/hooks/1//start/post/S31ganesha-start.sh)
changed: [
br014.bridges.psc.edu] =>
(item=/var/lib/glusterd/hooks/1/start/post/S29CTDBsetup.sh)
changed: [
br014.bridges.psc.edu] =>
(item=/var/lib/glusterd/hooks/1/stop/pre/S29CTDB-teardown.sh)
TASK [gluster.features/roles/gluster_hci : Check if valid FQDN is provided]
****
changed: [
br014.bridges.psc.edu -> localhost] => (
item=br014.bridges.psc.edu
)
TASK [gluster.features/roles/gluster_hci : Check if /var/log has enough
disk space] ***
fatal: [
br014.bridges.psc.edu]: FAILED! => {"changed": true, "cmd":
"df -m
/var/log | awk '/[0-9]%/ {print $4}'", "delta":
"0:00:00.008513", "end":
"2019-05-09 20:09:27.914400", "failed_when_result": true,
"rc": 0, "start":
"2019-05-09 20:09:27.905887", "stderr": "",
"stderr_lines": [], "stdout":
"7470", "stdout_lines": ["7470"]}
NO MORE HOSTS LEFT
*************************************************************
NO MORE HOSTS LEFT
*************************************************************
to retry, use: --limit
@/usr/share/cockpit/ovirt-dashboard/ansible/hc_wizard.retry
PLAY RECAP
*********************************************************************
br014.bridges.psc.edu : ok=25 changed=3 unreachable=0
failed=1