As title, we want to use 3 hosts with hyperconverged solution of oVirt. We installed oVirt
and Gluster as reported in the guide
https://www.ovirt.org/documentation/gluster-hyperconverged/chap-Deploying...
When we try to deploy, we get some errors which we cannot figure out.
==============================================================
gdeploy creates as configuration rules:
hc_nodes:
hosts:
virtnodetest-0-0:
gluster_infra_volume_groups:
- vgname: gluster_vg_sdb
pvname: /dev/sdb
gluster_infra_mount_devices:
- path: /gluster_bricks/engine
lvname: gluster_lv_engine
vgname: gluster_vg_sdb
- path: /gluster_bricks/isostorage
lvname: gluster_lv_isostorage
vgname: gluster_vg_sdb
- path: /gluster_bricks/vmstorage
lvname: gluster_lv_vmstorage
vgname: gluster_vg_sdb
gluster_infra_thick_lvs:
- vgname: gluster_vg_sdb
lvname: gluster_lv_engine
size: 150G
gluster_infra_thinpools:
- vgname: gluster_vg_sdb
thinpoolname: gluster_thinpool_gluster_vg_sdb
poolmetadatasize: 16G
gluster_infra_lv_logicalvols:
- vgname: gluster_vg_sdb
thinpool: gluster_thinpool_gluster_vg_sdb
lvname: gluster_lv_isostorage
lvsize: 250G
- vgname: gluster_vg_sdb
thinpool: gluster_thinpool_gluster_vg_sdb
lvname: gluster_lv_vmstorage
lvsize: 3500G
virtnodetest-0-1:
gluster_infra_volume_groups:
- vgname: gluster_vg_sdb
pvname: /dev/sdb
gluster_infra_mount_devices:
- path: /gluster_bricks/engine
lvname: gluster_lv_engine
vgname: gluster_vg_sdb
- path: /gluster_bricks/isostorage
lvname: gluster_lv_isostorage
vgname: gluster_vg_sdb
- path: /gluster_bricks/vmstorage
lvname: gluster_lv_vmstorage
vgname: gluster_vg_sdb
gluster_infra_thick_lvs:
- vgname: gluster_vg_sdb
lvname: gluster_lv_engine
size: 150G
gluster_infra_thinpools:
- vgname: gluster_vg_sdb
thinpoolname: gluster_thinpool_gluster_vg_sdb
poolmetadatasize: 16G
gluster_infra_lv_logicalvols:
- vgname: gluster_vg_sdb
thinpool: gluster_thinpool_gluster_vg_sdb
lvname: gluster_lv_isostorage
lvsize: 250G
- vgname: gluster_vg_sdb
thinpool: gluster_thinpool_gluster_vg_sdb
lvname: gluster_lv_vmstorage
lvsize: 3500G
virtnodetest-0-2:
gluster_infra_volume_groups:
- vgname: gluster_vg_sdb
pvname: /dev/sdb
gluster_infra_mount_devices:
- path: /gluster_bricks/engine
lvname: gluster_lv_engine
vgname: gluster_vg_sdb
- path: /gluster_bricks/isostorage
lvname: gluster_lv_isostorage
vgname: gluster_vg_sdb
- path: /gluster_bricks/vmstorage
lvname: gluster_lv_vmstorage
vgname: gluster_vg_sdb
gluster_infra_thick_lvs:
- vgname: gluster_vg_sdb
lvname: gluster_lv_engine
size: 150G
gluster_infra_thinpools:
- vgname: gluster_vg_sdb
thinpoolname: gluster_thinpool_gluster_vg_sdb
poolmetadatasize: 16G
gluster_infra_lv_logicalvols:
- vgname: gluster_vg_sdb
thinpool: gluster_thinpool_gluster_vg_sdb
lvname: gluster_lv_isostorage
lvsize: 250G
- vgname: gluster_vg_sdb
thinpool: gluster_thinpool_gluster_vg_sdb
lvname: gluster_lv_vmstorage
lvsize: 3500G
vars:
gluster_infra_disktype: JBOD
gluster_set_selinux_labels: true
gluster_infra_fw_ports:
- 2049/tcp
- 54321/tcp
- 5900/tcp
- 5900-6923/tcp
- 5666/tcp
- 16514/tcp
gluster_infra_fw_permanent: true
gluster_infra_fw_state: enabled
gluster_infra_fw_zone: public
gluster_infra_fw_services:
- glusterfs
gluster_features_force_varlogsizecheck: false
cluster_nodes:
- virtnodetest-0-0
- virtnodetest-0-1
- virtnodetest-0-2
gluster_features_hci_cluster: '{{ cluster_nodes }}'
gluster_features_hci_volumes:
- volname: engine
brick: /gluster_bricks/engine/engine
arbiter: 0
- volname: isostorage
brick: /gluster_bricks/isostorage/isostorage
arbiter: 0
- volname: vmstorage
brick: /gluster_bricks/vmstorage/vmstorage
arbiter: 0
=========================================================================
The system returns this error:
PLAY [Setup backend] ***********************************************************
TASK [Gathering Facts] *********************************************************
ok: [virtnodetest-0-2]
ok: [virtnodetest-0-1]
ok: [virtnodetest-0-0]
TASK [Check if valid hostnames are provided] ***********************************
changed: [virtnodetest-0-1] => (item=virtnodetest-0-1)
changed: [virtnodetest-0-1] => (item=virtnodetest-0-0)
changed: [virtnodetest-0-1] => (item=virtnodetest-0-2)
TASK [Check if provided hostnames are valid] ***********************************
ok: [virtnodetest-0-1] => {
"changed": false,
"msg": "All assertions passed"
}
ok: [virtnodetest-0-0] => {
"changed": false,
"msg": "All assertions passed"
}
ok: [virtnodetest-0-2] => {
"changed": false,
"msg": "All assertions passed"
}
TASK [Check if /var/log has enough disk space] *********************************
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]
TASK [Check if the /var is greater than 15G] ***********************************
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]
TASK [Check if disks have logical block size of 512B] **************************
skipping: [virtnodetest-0-1] => (item={u'vgname': u'gluster_vg_sdb',
u'pvname': u'/dev/sdb'})
skipping: [virtnodetest-0-0] => (item={u'vgname': u'gluster_vg_sdb',
u'pvname': u'/dev/sdb'})
skipping: [virtnodetest-0-2] => (item={u'vgname': u'gluster_vg_sdb',
u'pvname': u'/dev/sdb'})
TASK [Check if logical block size is 512 bytes] ********************************
skipping: [virtnodetest-0-1] => (item=Logical Block Size)
skipping: [virtnodetest-0-0] => (item=Logical Block Size)
skipping: [virtnodetest-0-2] => (item=Logical Block Size)
TASK [Get logical block size of VDO devices] ***********************************
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]
TASK [Check if logical block size is 512 bytes for VDO devices] ****************
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/firewall_config : Start firewalld if not already started] ***
ok: [virtnodetest-0-2]
ok: [virtnodetest-0-1]
ok: [virtnodetest-0-0]
TASK [gluster.infra/roles/firewall_config : check if required variables are set] ***
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/firewall_config : Open/Close firewalld ports] ********
ok: [virtnodetest-0-2] => (item=2049/tcp)
ok: [virtnodetest-0-1] => (item=2049/tcp)
ok: [virtnodetest-0-0] => (item=2049/tcp)
ok: [virtnodetest-0-2] => (item=54321/tcp)
ok: [virtnodetest-0-0] => (item=54321/tcp)
ok: [virtnodetest-0-1] => (item=54321/tcp)
ok: [virtnodetest-0-2] => (item=5900/tcp)
ok: [virtnodetest-0-1] => (item=5900/tcp)
ok: [virtnodetest-0-0] => (item=5900/tcp)
ok: [virtnodetest-0-2] => (item=5900-6923/tcp)
ok: [virtnodetest-0-0] => (item=5900-6923/tcp)
ok: [virtnodetest-0-1] => (item=5900-6923/tcp)
ok: [virtnodetest-0-2] => (item=5666/tcp)
ok: [virtnodetest-0-1] => (item=5666/tcp)
ok: [virtnodetest-0-0] => (item=5666/tcp)
ok: [virtnodetest-0-2] => (item=16514/tcp)
ok: [virtnodetest-0-1] => (item=16514/tcp)
ok: [virtnodetest-0-0] => (item=16514/tcp)
TASK [gluster.infra/roles/firewall_config : Add/Delete services to firewalld rules] ***
ok: [virtnodetest-0-1] => (item=glusterfs)
ok: [virtnodetest-0-0] => (item=glusterfs)
ok: [virtnodetest-0-2] => (item=glusterfs)
TASK [gluster.infra/roles/backend_setup : Check if vdsm-python package is installed or
not] ***
changed: [virtnodetest-0-2]
changed: [virtnodetest-0-1]
changed: [virtnodetest-0-0]
TASK [gluster.infra/roles/backend_setup : Remove the existing LVM filter] ******
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Check that the multipath.conf exists] ***
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Ensure that multipathd services is enabled if
not] ***
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Ensure that multipathd services is running] ***
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Create /etc/multipath/conf.d if doesn't
exists] ***
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Get the UUID of the devices] *********
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Check that the blacklist.conf exists] ***
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Create blacklist template content] ***
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Add wwid to blacklist in blacklist.conf file]
***
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Reload multipathd] *******************
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Gather facts to determine the OS distribution]
***
ok: [virtnodetest-0-1]
ok: [virtnodetest-0-0]
ok: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Change to Install lvm tools for debian systems.]
***
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Change to Install lvm tools for RHEL systems.]
***
ok: [virtnodetest-0-2]
ok: [virtnodetest-0-1]
ok: [virtnodetest-0-0]
TASK [gluster.infra/roles/backend_setup : Install python-yaml package for Debian systems]
***
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Initialize vdo_devs array] ***********
ok: [virtnodetest-0-1]
ok: [virtnodetest-0-0]
ok: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Record VDO devices (if any)] *********
skipping: [virtnodetest-0-1] => (item={u'vgname': u'gluster_vg_sdb',
u'pvname': u'/dev/sdb'})
skipping: [virtnodetest-0-0] => (item={u'vgname': u'gluster_vg_sdb',
u'pvname': u'/dev/sdb'})
skipping: [virtnodetest-0-2] => (item={u'vgname': u'gluster_vg_sdb',
u'pvname': u'/dev/sdb'})
TASK [gluster.infra/roles/backend_setup : Configure lvm thinpool extend threshold] ***
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Configure lvm thinpool extend percentage] ***
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Check if vdo block device exists] ****
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Record for missing devices for phase 2] ***
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : set fact if it will at least install 1 vdo
device] ***
TASK [gluster.infra/roles/backend_setup : Install VDO dependencies] ************
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : set fact about vdo installed deps] ***
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Enable and start vdo service] ********
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Create VDO with specified size] ******
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Set VDO maxDiscardSize as 16M] *******
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Stop VDO volumes] ********************
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Start VDO volumes] *******************
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Check if valid disktype is provided] ***
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Set PV data alignment for JBOD] ******
ok: [virtnodetest-0-1]
ok: [virtnodetest-0-0]
ok: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Set PV data alignment for RAID] ******
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Set VG physical extent size for RAID] ***
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : include_tasks] ***********************
included: /etc/ansible/roles/gluster.infra/roles/backend_setup/tasks/get_vg_groupings.yml
for virtnodetest-0-1, virtnodetest-0-0, virtnodetest-0-2
TASK [gluster.infra/roles/backend_setup : Group devices by volume group name, including
existing devices] ***
ok: [virtnodetest-0-1]
ok: [virtnodetest-0-0]
ok: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Check if vg block device exists] *****
changed: [virtnodetest-0-0] => (item={u'key': u'gluster_vg_sdb',
u'value': [{u'vgname': u'gluster_vg_sdb', u'pvname':
u'/dev/sdb'}]})
changed: [virtnodetest-0-1] => (item={u'key': u'gluster_vg_sdb',
u'value': [{u'vgname': u'gluster_vg_sdb', u'pvname':
u'/dev/sdb'}]})
changed: [virtnodetest-0-2] => (item={u'key': u'gluster_vg_sdb',
u'value': [{u'vgname': u'gluster_vg_sdb', u'pvname':
u'/dev/sdb'}]})
TASK [gluster.infra/roles/backend_setup : Filter none-existing devices] ********
ok: [virtnodetest-0-1]
ok: [virtnodetest-0-0]
ok: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Record for missing devices for phase 2] ***
ok: [virtnodetest-0-1] => (item={u'stderr_lines': [],
u'ansible_loop_var': u'item', u'end': u'2021-05-13
12:18:33.575598', u'stderr': u'', u'stdout': u'0',
u'changed': True, u'failed': False, u'delta':
u'0:00:00.009901', u'cmd': u' test -b /dev/sdb && echo
"1" || echo "0"; \n', u'item': {u'value':
[{u'vgname': u'gluster_vg_sdb', u'pvname': u'/dev/sdb'}],
u'key': u'gluster_vg_sdb'}, u'rc': 0, u'invocation':
{u'module_args': {u'warn': True, u'executable': None,
u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params':
u' test -b /dev/sdb && echo "1" || echo "0"; \n',
u'removes': None, u'argv': None, u'creates': None,
u'chdir': None, u'stdin_add_newline': True, u'stdin': None}},
u'stdout_lines': [u'0'], u'start': u'2021-05-13
12:18:33.565697'})
ok: [virtnodetest-0-0] => (item={u'stderr_lines': [],
u'ansible_loop_var': u'item', u'end': u'2021-05-13
10:52:56.886693', u'stderr': u'', u'stdout': u'0',
u'changed': True, u'failed': False, u'delta':
u'0:00:00.008123', u'cmd': u' test -b /dev/sdb && echo
"1" || echo "0"; \n', u'item': {u'value':
[{u'vgname': u'gluster_vg_sdb', u'pvname': u'/dev/sdb'}],
u'key': u'gluster_vg_sdb'}, u'rc': 0, u'invocation':
{u'module_args': {u'warn': True, u'executable': None,
u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params':
u' test -b /dev/sdb && echo "1" || echo "0"; \n',
u'removes': None, u'argv': None, u'creates': None,
u'chdir': None, u'stdin_add_newline': True, u'stdin': None}},
u'stdout_lines': [u'0'], u'start': u'2021-05-13
10:52:56.878570'})
ok: [virtnodetest-0-2] => (item={u'stderr_lines': [],
u'ansible_loop_var': u'item', u'end': u'2021-05-13
12:25:24.420710', u'stderr': u'', u'stdout': u'0',
u'changed': True, u'failed': False, u'delta':
u'0:00:00.007307', u'cmd': u' test -b /dev/sdb && echo
"1" || echo "0"; \n', u'item': {u'value':
[{u'vgname': u'gluster_vg_sdb', u'pvname': u'/dev/sdb'}],
u'key': u'gluster_vg_sdb'}, u'rc': 0, u'invocation':
{u'module_args': {u'warn': True, u'executable': None,
u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params':
u' test -b /dev/sdb && echo "1" || echo "0"; \n',
u'removes': None, u'argv': None, u'creates': None,
u'chdir': None, u'stdin_add_newline': True, u'stdin': None}},
u'stdout_lines': [u'0'], u'start': u'2021-05-13
12:25:24.413403'})
TASK [gluster.infra/roles/backend_setup : Create volume groups] ****************
skipping: [virtnodetest-0-1] => (item={u'key': u'gluster_vg_sdb',
u'value': []})
skipping: [virtnodetest-0-0] => (item={u'key': u'gluster_vg_sdb',
u'value': []})
skipping: [virtnodetest-0-2] => (item={u'key': u'gluster_vg_sdb',
u'value': []})
TASK [gluster.infra/roles/backend_setup : update LVM fact's] *******************
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Check if thick-lv block devices exists] ***
changed: [virtnodetest-0-0] => (item={u'lvname': u'gluster_lv_engine',
u'vgname': u'gluster_vg_sdb', u'size': u'150G'})
changed: [virtnodetest-0-1] => (item={u'lvname': u'gluster_lv_engine',
u'vgname': u'gluster_vg_sdb', u'size': u'150G'})
changed: [virtnodetest-0-2] => (item={u'lvname': u'gluster_lv_engine',
u'vgname': u'gluster_vg_sdb', u'size': u'150G'})
TASK [gluster.infra/roles/backend_setup : Record for missing devices for phase 2] ***
skipping: [virtnodetest-0-1] => (item={u'stderr_lines': [],
u'ansible_loop_var': u'item', u'end': u'2021-05-13
12:18:37.528159', u'stderr': u'', u'stdout': u'1',
u'changed': True, u'failed': False, u'delta':
u'0:00:00.010032', u'cmd': u' echo "1" \n',
u'item': {u'lvname': u'gluster_lv_engine', u'vgname':
u'gluster_vg_sdb', u'size': u'150G'}, u'rc': 0,
u'invocation': {u'module_args': {u'warn': True,
u'executable': None, u'_uses_shell': True, u'strip_empty_ends':
True, u'_raw_params': u' echo "1" \n', u'removes': None,
u'argv': None, u'creates': None, u'chdir': None,
u'stdin_add_newline': True, u'stdin': None}}, u'stdout_lines':
[u'1'], u'start': u'2021-05-13 12:18:37.518127'})
skipping: [virtnodetest-0-0] => (item={u'stderr_lines': [],
u'ansible_loop_var': u'item', u'end': u'2021-05-13
10:53:00.863436', u'stderr': u'', u'stdout': u'1',
u'changed': True, u'failed': False, u'delta':
u'0:00:00.007459', u'cmd': u' echo "1" \n',
u'item': {u'lvname': u'gluster_lv_engine', u'vgname':
u'gluster_vg_sdb', u'size': u'150G'}, u'rc': 0,
u'invocation': {u'module_args': {u'warn': True,
u'executable': None, u'_uses_shell': True, u'strip_empty_ends':
True, u'_raw_params': u' echo "1" \n', u'removes': None,
u'argv': None, u'creates': None, u'chdir': None,
u'stdin_add_newline': True, u'stdin': None}}, u'stdout_lines':
[u'1'], u'start': u'2021-05-13 10:53:00.855977'})
skipping: [virtnodetest-0-2] => (item={u'stderr_lines': [],
u'ansible_loop_var': u'item', u'end': u'2021-05-13
12:25:28.261106', u'stderr': u'', u'stdout': u'1',
u'changed': True, u'failed': False, u'delta':
u'0:00:00.007818', u'cmd': u' echo "1" \n',
u'item': {u'lvname': u'gluster_lv_engine', u'vgname':
u'gluster_vg_sdb', u'size': u'150G'}, u'rc': 0,
u'invocation': {u'module_args': {u'warn': True,
u'executable': None, u'_uses_shell': True, u'strip_empty_ends':
True, u'_raw_params': u' echo "1" \n', u'removes': None,
u'argv': None, u'creates': None, u'chdir': None,
u'stdin_add_newline': True, u'stdin': None}}, u'stdout_lines':
[u'1'], u'start': u'2021-05-13 12:25:28.253288'})
TASK [gluster.infra/roles/backend_setup : include_tasks] ***********************
included: /etc/ansible/roles/gluster.infra/roles/backend_setup/tasks/get_vg_groupings.yml
for virtnodetest-0-1, virtnodetest-0-0, virtnodetest-0-2
TASK [gluster.infra/roles/backend_setup : Group devices by volume group name, including
existing devices] ***
ok: [virtnodetest-0-1]
ok: [virtnodetest-0-0]
ok: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Check if vg block device exists] *****
TASK [gluster.infra/roles/backend_setup : Filter none-existing devices] ********
ok: [virtnodetest-0-1]
ok: [virtnodetest-0-0]
ok: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Make sure thick pvs exists in volume group] ***
TASK [gluster.infra/roles/backend_setup : update LVM fact's] *******************
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Create thick logical volume] *********
failed: [virtnodetest-0-1] (item={u'lvname': u'gluster_lv_engine',
u'vgname': u'gluster_vg_sdb', u'size': u'150G'}) =>
{"ansible_index_var": "index", "ansible_loop_var":
"item", "changed": false, "err": " WARNING: Device for
PV gx6iUE-369Z-3FDP-aRUQ-Wur0-1Xhf-v4g79j not found or rejected by a filter.\n Volume
group \"gluster_vg_sdb\" not found\n Cannot process volume group
gluster_vg_sdb\n", "index": 0, "item": {"lvname":
"gluster_lv_engine", "size": "150G", "vgname":
"gluster_vg_sdb"}, "msg": "Volume group gluster_vg_sdb does not
exist.", "rc": 5}
failed: [virtnodetest-0-0] (item={u'lvname': u'gluster_lv_engine',
u'vgname': u'gluster_vg_sdb', u'size': u'150G'}) =>
{"ansible_index_var": "index", "ansible_loop_var":
"item", "changed": false, "err": " Volume group
\"gluster_vg_sdb\" not found\n Cannot process volume group
gluster_vg_sdb\n", "index": 0, "item": {"lvname":
"gluster_lv_engine", "size": "150G", "vgname":
"gluster_vg_sdb"}, "msg": "Volume group gluster_vg_sdb does not
exist.", "rc": 5}
failed: [virtnodetest-0-2] (item={u'lvname': u'gluster_lv_engine',
u'vgname': u'gluster_vg_sdb', u'size': u'150G'}) =>
{"ansible_index_var": "index", "ansible_loop_var":
"item", "changed": false, "err": " Volume group
\"gluster_vg_sdb\" not found\n Cannot process volume group
gluster_vg_sdb\n", "index": 0, "item": {"lvname":
"gluster_lv_engine", "size": "150G", "vgname":
"gluster_vg_sdb"}, "msg": "Volume group gluster_vg_sdb does not
exist.", "rc": 5}
NO MORE HOSTS LEFT *************************************************************
NO MORE HOSTS LEFT *************************************************************
PLAY RECAP *********************************************************************
virtnodetest-0-0 : ok=19 changed=3 unreachable=0 failed=1 skipped=41
rescued=0 ignored=0
virtnodetest-0-1 : ok=20 changed=4 unreachable=0 failed=1 skipped=41
rescued=0 ignored=0
virtnodetest-0-2 : ok=19 changed=3 unreachable=0 failed=1 skipped=41
rescued=0 ignored=0
Please check /var/log/cockpit/ovirt-dashboard/gluster-deployment.log for more
informations.
======================================================================
How can we resolve this issue?