
As title, we want to use 3 hosts with hyperconverged solution of oVirt. We installed oVirt and Gluster as reported in the guide https://www.ovirt.org/documentation/gluster-hyperconverged/chap-Deploying_Hy... When we try to deploy, we get some errors which we cannot figure out. ============================================================== gdeploy creates as configuration rules: hc_nodes: hosts: virtnodetest-0-0: gluster_infra_volume_groups: - vgname: gluster_vg_sdb pvname: /dev/sdb gluster_infra_mount_devices: - path: /gluster_bricks/engine lvname: gluster_lv_engine vgname: gluster_vg_sdb - path: /gluster_bricks/isostorage lvname: gluster_lv_isostorage vgname: gluster_vg_sdb - path: /gluster_bricks/vmstorage lvname: gluster_lv_vmstorage vgname: gluster_vg_sdb gluster_infra_thick_lvs: - vgname: gluster_vg_sdb lvname: gluster_lv_engine size: 150G gluster_infra_thinpools: - vgname: gluster_vg_sdb thinpoolname: gluster_thinpool_gluster_vg_sdb poolmetadatasize: 16G gluster_infra_lv_logicalvols: - vgname: gluster_vg_sdb thinpool: gluster_thinpool_gluster_vg_sdb lvname: gluster_lv_isostorage lvsize: 250G - vgname: gluster_vg_sdb thinpool: gluster_thinpool_gluster_vg_sdb lvname: gluster_lv_vmstorage lvsize: 3500G virtnodetest-0-1: gluster_infra_volume_groups: - vgname: gluster_vg_sdb pvname: /dev/sdb gluster_infra_mount_devices: - path: /gluster_bricks/engine lvname: gluster_lv_engine vgname: gluster_vg_sdb - path: /gluster_bricks/isostorage lvname: gluster_lv_isostorage vgname: gluster_vg_sdb - path: /gluster_bricks/vmstorage lvname: gluster_lv_vmstorage vgname: gluster_vg_sdb gluster_infra_thick_lvs: - vgname: gluster_vg_sdb lvname: gluster_lv_engine size: 150G gluster_infra_thinpools: - vgname: gluster_vg_sdb thinpoolname: gluster_thinpool_gluster_vg_sdb poolmetadatasize: 16G gluster_infra_lv_logicalvols: - vgname: gluster_vg_sdb thinpool: gluster_thinpool_gluster_vg_sdb lvname: gluster_lv_isostorage lvsize: 250G - vgname: gluster_vg_sdb thinpool: gluster_thinpool_gluster_vg_sdb lvname: gluster_lv_vmstorage lvsize: 3500G virtnodetest-0-2: gluster_infra_volume_groups: - vgname: gluster_vg_sdb pvname: /dev/sdb gluster_infra_mount_devices: - path: /gluster_bricks/engine lvname: gluster_lv_engine vgname: gluster_vg_sdb - path: /gluster_bricks/isostorage lvname: gluster_lv_isostorage vgname: gluster_vg_sdb - path: /gluster_bricks/vmstorage lvname: gluster_lv_vmstorage vgname: gluster_vg_sdb gluster_infra_thick_lvs: - vgname: gluster_vg_sdb lvname: gluster_lv_engine size: 150G gluster_infra_thinpools: - vgname: gluster_vg_sdb thinpoolname: gluster_thinpool_gluster_vg_sdb poolmetadatasize: 16G gluster_infra_lv_logicalvols: - vgname: gluster_vg_sdb thinpool: gluster_thinpool_gluster_vg_sdb lvname: gluster_lv_isostorage lvsize: 250G - vgname: gluster_vg_sdb thinpool: gluster_thinpool_gluster_vg_sdb lvname: gluster_lv_vmstorage lvsize: 3500G vars: gluster_infra_disktype: JBOD gluster_set_selinux_labels: true gluster_infra_fw_ports: - 2049/tcp - 54321/tcp - 5900/tcp - 5900-6923/tcp - 5666/tcp - 16514/tcp gluster_infra_fw_permanent: true gluster_infra_fw_state: enabled gluster_infra_fw_zone: public gluster_infra_fw_services: - glusterfs gluster_features_force_varlogsizecheck: false cluster_nodes: - virtnodetest-0-0 - virtnodetest-0-1 - virtnodetest-0-2 gluster_features_hci_cluster: '{{ cluster_nodes }}' gluster_features_hci_volumes: - volname: engine brick: /gluster_bricks/engine/engine arbiter: 0 - volname: isostorage brick: /gluster_bricks/isostorage/isostorage arbiter: 0 - volname: vmstorage brick: /gluster_bricks/vmstorage/vmstorage arbiter: 0 ========================================================================= The system returns this error: PLAY [Setup backend] *********************************************************** TASK [Gathering Facts] ********************************************************* ok: [virtnodetest-0-2] ok: [virtnodetest-0-1] ok: [virtnodetest-0-0] TASK [Check if valid hostnames are provided] *********************************** changed: [virtnodetest-0-1] => (item=virtnodetest-0-1) changed: [virtnodetest-0-1] => (item=virtnodetest-0-0) changed: [virtnodetest-0-1] => (item=virtnodetest-0-2) TASK [Check if provided hostnames are valid] *********************************** ok: [virtnodetest-0-1] => { "changed": false, "msg": "All assertions passed" } ok: [virtnodetest-0-0] => { "changed": false, "msg": "All assertions passed" } ok: [virtnodetest-0-2] => { "changed": false, "msg": "All assertions passed" } TASK [Check if /var/log has enough disk space] ********************************* skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2] TASK [Check if the /var is greater than 15G] *********************************** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2] TASK [Check if disks have logical block size of 512B] ************************** skipping: [virtnodetest-0-1] => (item={u'vgname': u'gluster_vg_sdb', u'pvname': u'/dev/sdb'}) skipping: [virtnodetest-0-0] => (item={u'vgname': u'gluster_vg_sdb', u'pvname': u'/dev/sdb'}) skipping: [virtnodetest-0-2] => (item={u'vgname': u'gluster_vg_sdb', u'pvname': u'/dev/sdb'}) TASK [Check if logical block size is 512 bytes] ******************************** skipping: [virtnodetest-0-1] => (item=Logical Block Size) skipping: [virtnodetest-0-0] => (item=Logical Block Size) skipping: [virtnodetest-0-2] => (item=Logical Block Size) TASK [Get logical block size of VDO devices] *********************************** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2] TASK [Check if logical block size is 512 bytes for VDO devices] **************** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2] TASK [gluster.infra/roles/firewall_config : Start firewalld if not already started] *** ok: [virtnodetest-0-2] ok: [virtnodetest-0-1] ok: [virtnodetest-0-0] TASK [gluster.infra/roles/firewall_config : check if required variables are set] *** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2] TASK [gluster.infra/roles/firewall_config : Open/Close firewalld ports] ******** ok: [virtnodetest-0-2] => (item=2049/tcp) ok: [virtnodetest-0-1] => (item=2049/tcp) ok: [virtnodetest-0-0] => (item=2049/tcp) ok: [virtnodetest-0-2] => (item=54321/tcp) ok: [virtnodetest-0-0] => (item=54321/tcp) ok: [virtnodetest-0-1] => (item=54321/tcp) ok: [virtnodetest-0-2] => (item=5900/tcp) ok: [virtnodetest-0-1] => (item=5900/tcp) ok: [virtnodetest-0-0] => (item=5900/tcp) ok: [virtnodetest-0-2] => (item=5900-6923/tcp) ok: [virtnodetest-0-0] => (item=5900-6923/tcp) ok: [virtnodetest-0-1] => (item=5900-6923/tcp) ok: [virtnodetest-0-2] => (item=5666/tcp) ok: [virtnodetest-0-1] => (item=5666/tcp) ok: [virtnodetest-0-0] => (item=5666/tcp) ok: [virtnodetest-0-2] => (item=16514/tcp) ok: [virtnodetest-0-1] => (item=16514/tcp) ok: [virtnodetest-0-0] => (item=16514/tcp) TASK [gluster.infra/roles/firewall_config : Add/Delete services to firewalld rules] *** ok: [virtnodetest-0-1] => (item=glusterfs) ok: [virtnodetest-0-0] => (item=glusterfs) ok: [virtnodetest-0-2] => (item=glusterfs) TASK [gluster.infra/roles/backend_setup : Check if vdsm-python package is installed or not] *** changed: [virtnodetest-0-2] changed: [virtnodetest-0-1] changed: [virtnodetest-0-0] TASK [gluster.infra/roles/backend_setup : Remove the existing LVM filter] ****** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2] TASK [gluster.infra/roles/backend_setup : Check that the multipath.conf exists] *** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2] TASK [gluster.infra/roles/backend_setup : Ensure that multipathd services is enabled if not] *** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2] TASK [gluster.infra/roles/backend_setup : Ensure that multipathd services is running] *** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2] TASK [gluster.infra/roles/backend_setup : Create /etc/multipath/conf.d if doesn't exists] *** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2] TASK [gluster.infra/roles/backend_setup : Get the UUID of the devices] ********* skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2] TASK [gluster.infra/roles/backend_setup : Check that the blacklist.conf exists] *** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2] TASK [gluster.infra/roles/backend_setup : Create blacklist template content] *** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2] TASK [gluster.infra/roles/backend_setup : Add wwid to blacklist in blacklist.conf file] *** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2] TASK [gluster.infra/roles/backend_setup : Reload multipathd] ******************* skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2] TASK [gluster.infra/roles/backend_setup : Gather facts to determine the OS distribution] *** ok: [virtnodetest-0-1] ok: [virtnodetest-0-0] ok: [virtnodetest-0-2] TASK [gluster.infra/roles/backend_setup : Change to Install lvm tools for debian systems.] *** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2] TASK [gluster.infra/roles/backend_setup : Change to Install lvm tools for RHEL systems.] *** ok: [virtnodetest-0-2] ok: [virtnodetest-0-1] ok: [virtnodetest-0-0] TASK [gluster.infra/roles/backend_setup : Install python-yaml package for Debian systems] *** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2] TASK [gluster.infra/roles/backend_setup : Initialize vdo_devs array] *********** ok: [virtnodetest-0-1] ok: [virtnodetest-0-0] ok: [virtnodetest-0-2] TASK [gluster.infra/roles/backend_setup : Record VDO devices (if any)] ********* skipping: [virtnodetest-0-1] => (item={u'vgname': u'gluster_vg_sdb', u'pvname': u'/dev/sdb'}) skipping: [virtnodetest-0-0] => (item={u'vgname': u'gluster_vg_sdb', u'pvname': u'/dev/sdb'}) skipping: [virtnodetest-0-2] => (item={u'vgname': u'gluster_vg_sdb', u'pvname': u'/dev/sdb'}) TASK [gluster.infra/roles/backend_setup : Configure lvm thinpool extend threshold] *** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2] TASK [gluster.infra/roles/backend_setup : Configure lvm thinpool extend percentage] *** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2] TASK [gluster.infra/roles/backend_setup : Check if vdo block device exists] **** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2] TASK [gluster.infra/roles/backend_setup : Record for missing devices for phase 2] *** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2] TASK [gluster.infra/roles/backend_setup : set fact if it will at least install 1 vdo device] *** TASK [gluster.infra/roles/backend_setup : Install VDO dependencies] ************ skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2] TASK [gluster.infra/roles/backend_setup : set fact about vdo installed deps] *** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2] TASK [gluster.infra/roles/backend_setup : Enable and start vdo service] ******** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2] TASK [gluster.infra/roles/backend_setup : Create VDO with specified size] ****** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2] TASK [gluster.infra/roles/backend_setup : Set VDO maxDiscardSize as 16M] ******* skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2] TASK [gluster.infra/roles/backend_setup : Stop VDO volumes] ******************** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2] TASK [gluster.infra/roles/backend_setup : Start VDO volumes] ******************* skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2] TASK [gluster.infra/roles/backend_setup : Check if valid disktype is provided] *** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2] TASK [gluster.infra/roles/backend_setup : Set PV data alignment for JBOD] ****** ok: [virtnodetest-0-1] ok: [virtnodetest-0-0] ok: [virtnodetest-0-2] TASK [gluster.infra/roles/backend_setup : Set PV data alignment for RAID] ****** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2] TASK [gluster.infra/roles/backend_setup : Set VG physical extent size for RAID] *** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2] TASK [gluster.infra/roles/backend_setup : include_tasks] *********************** included: /etc/ansible/roles/gluster.infra/roles/backend_setup/tasks/get_vg_groupings.yml for virtnodetest-0-1, virtnodetest-0-0, virtnodetest-0-2 TASK [gluster.infra/roles/backend_setup : Group devices by volume group name, including existing devices] *** ok: [virtnodetest-0-1] ok: [virtnodetest-0-0] ok: [virtnodetest-0-2] TASK [gluster.infra/roles/backend_setup : Check if vg block device exists] ***** changed: [virtnodetest-0-0] => (item={u'key': u'gluster_vg_sdb', u'value': [{u'vgname': u'gluster_vg_sdb', u'pvname': u'/dev/sdb'}]}) changed: [virtnodetest-0-1] => (item={u'key': u'gluster_vg_sdb', u'value': [{u'vgname': u'gluster_vg_sdb', u'pvname': u'/dev/sdb'}]}) changed: [virtnodetest-0-2] => (item={u'key': u'gluster_vg_sdb', u'value': [{u'vgname': u'gluster_vg_sdb', u'pvname': u'/dev/sdb'}]}) TASK [gluster.infra/roles/backend_setup : Filter none-existing devices] ******** ok: [virtnodetest-0-1] ok: [virtnodetest-0-0] ok: [virtnodetest-0-2] TASK [gluster.infra/roles/backend_setup : Record for missing devices for phase 2] *** ok: [virtnodetest-0-1] => (item={u'stderr_lines': [], u'ansible_loop_var': u'item', u'end': u'2021-05-13 12:18:33.575598', u'stderr': u'', u'stdout': u'0', u'changed': True, u'failed': False, u'delta': u'0:00:00.009901', u'cmd': u' test -b /dev/sdb && echo "1" || echo "0"; \n', u'item': {u'value': [{u'vgname': u'gluster_vg_sdb', u'pvname': u'/dev/sdb'}], u'key': u'gluster_vg_sdb'}, u'rc': 0, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u' test -b /dev/sdb && echo "1" || echo "0"; \n', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'stdout_lines': [u'0'], u'start': u'2021-05-13 12:18:33.565697'}) ok: [virtnodetest-0-0] => (item={u'stderr_lines': [], u'ansible_loop_var': u'item', u'end': u'2021-05-13 10:52:56.886693', u'stderr': u'', u'stdout': u'0', u'changed': True, u'failed': False, u'delta': u'0:00:00.008123', u'cmd': u' test -b /dev/sdb && echo "1" || echo "0"; \n', u'item': {u'value': [{u'vgname': u'gluster_vg_sdb', u'pvname': u'/dev/sdb'}], u'key': u'gluster_vg_sdb'}, u'rc': 0, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u' test -b /dev/sdb && echo "1" || echo "0"; \n', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'stdout_lines': [u'0'], u'start': u'2021-05-13 10:52:56.878570'}) ok: [virtnodetest-0-2] => (item={u'stderr_lines': [], u'ansible_loop_var': u'item', u'end': u'2021-05-13 12:25:24.420710', u'stderr': u'', u'stdout': u'0', u'changed': True, u'failed': False, u'delta': u'0:00:00.007307', u'cmd': u' test -b /dev/sdb && echo "1" || echo "0"; \n', u'item': {u'value': [{u'vgname': u'gluster_vg_sdb', u'pvname': u'/dev/sdb'}], u'key': u'gluster_vg_sdb'}, u'rc': 0, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u' test -b /dev/sdb && echo "1" || echo "0"; \n', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'stdout_lines': [u'0'], u'start': u'2021-05-13 12:25:24.413403'}) TASK [gluster.infra/roles/backend_setup : Create volume groups] **************** skipping: [virtnodetest-0-1] => (item={u'key': u'gluster_vg_sdb', u'value': []}) skipping: [virtnodetest-0-0] => (item={u'key': u'gluster_vg_sdb', u'value': []}) skipping: [virtnodetest-0-2] => (item={u'key': u'gluster_vg_sdb', u'value': []}) TASK [gluster.infra/roles/backend_setup : update LVM fact's] ******************* skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2] TASK [gluster.infra/roles/backend_setup : Check if thick-lv block devices exists] *** changed: [virtnodetest-0-0] => (item={u'lvname': u'gluster_lv_engine', u'vgname': u'gluster_vg_sdb', u'size': u'150G'}) changed: [virtnodetest-0-1] => (item={u'lvname': u'gluster_lv_engine', u'vgname': u'gluster_vg_sdb', u'size': u'150G'}) changed: [virtnodetest-0-2] => (item={u'lvname': u'gluster_lv_engine', u'vgname': u'gluster_vg_sdb', u'size': u'150G'}) TASK [gluster.infra/roles/backend_setup : Record for missing devices for phase 2] *** skipping: [virtnodetest-0-1] => (item={u'stderr_lines': [], u'ansible_loop_var': u'item', u'end': u'2021-05-13 12:18:37.528159', u'stderr': u'', u'stdout': u'1', u'changed': True, u'failed': False, u'delta': u'0:00:00.010032', u'cmd': u' echo "1" \n', u'item': {u'lvname': u'gluster_lv_engine', u'vgname': u'gluster_vg_sdb', u'size': u'150G'}, u'rc': 0, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u' echo "1" \n', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'stdout_lines': [u'1'], u'start': u'2021-05-13 12:18:37.518127'}) skipping: [virtnodetest-0-0] => (item={u'stderr_lines': [], u'ansible_loop_var': u'item', u'end': u'2021-05-13 10:53:00.863436', u'stderr': u'', u'stdout': u'1', u'changed': True, u'failed': False, u'delta': u'0:00:00.007459', u'cmd': u' echo "1" \n', u'item': {u'lvname': u'gluster_lv_engine', u'vgname': u'gluster_vg_sdb', u'size': u'150G'}, u'rc': 0, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u' echo "1" \n', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'stdout_lines': [u'1'], u'start': u'2021-05-13 10:53:00.855977'}) skipping: [virtnodetest-0-2] => (item={u'stderr_lines': [], u'ansible_loop_var': u'item', u'end': u'2021-05-13 12:25:28.261106', u'stderr': u'', u'stdout': u'1', u'changed': True, u'failed': False, u'delta': u'0:00:00.007818', u'cmd': u' echo "1" \n', u'item': {u'lvname': u'gluster_lv_engine', u'vgname': u'gluster_vg_sdb', u'size': u'150G'}, u'rc': 0, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u' echo "1" \n', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'stdout_lines': [u'1'], u'start': u'2021-05-13 12:25:28.253288'}) TASK [gluster.infra/roles/backend_setup : include_tasks] *********************** included: /etc/ansible/roles/gluster.infra/roles/backend_setup/tasks/get_vg_groupings.yml for virtnodetest-0-1, virtnodetest-0-0, virtnodetest-0-2 TASK [gluster.infra/roles/backend_setup : Group devices by volume group name, including existing devices] *** ok: [virtnodetest-0-1] ok: [virtnodetest-0-0] ok: [virtnodetest-0-2] TASK [gluster.infra/roles/backend_setup : Check if vg block device exists] ***** TASK [gluster.infra/roles/backend_setup : Filter none-existing devices] ******** ok: [virtnodetest-0-1] ok: [virtnodetest-0-0] ok: [virtnodetest-0-2] TASK [gluster.infra/roles/backend_setup : Make sure thick pvs exists in volume group] *** TASK [gluster.infra/roles/backend_setup : update LVM fact's] ******************* skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2] TASK [gluster.infra/roles/backend_setup : Create thick logical volume] ********* failed: [virtnodetest-0-1] (item={u'lvname': u'gluster_lv_engine', u'vgname': u'gluster_vg_sdb', u'size': u'150G'}) => {"ansible_index_var": "index", "ansible_loop_var": "item", "changed": false, "err": " WARNING: Device for PV gx6iUE-369Z-3FDP-aRUQ-Wur0-1Xhf-v4g79j not found or rejected by a filter.\n Volume group \"gluster_vg_sdb\" not found\n Cannot process volume group gluster_vg_sdb\n", "index": 0, "item": {"lvname": "gluster_lv_engine", "size": "150G", "vgname": "gluster_vg_sdb"}, "msg": "Volume group gluster_vg_sdb does not exist.", "rc": 5} failed: [virtnodetest-0-0] (item={u'lvname': u'gluster_lv_engine', u'vgname': u'gluster_vg_sdb', u'size': u'150G'}) => {"ansible_index_var": "index", "ansible_loop_var": "item", "changed": false, "err": " Volume group \"gluster_vg_sdb\" not found\n Cannot process volume group gluster_vg_sdb\n", "index": 0, "item": {"lvname": "gluster_lv_engine", "size": "150G", "vgname": "gluster_vg_sdb"}, "msg": "Volume group gluster_vg_sdb does not exist.", "rc": 5} failed: [virtnodetest-0-2] (item={u'lvname': u'gluster_lv_engine', u'vgname': u'gluster_vg_sdb', u'size': u'150G'}) => {"ansible_index_var": "index", "ansible_loop_var": "item", "changed": false, "err": " Volume group \"gluster_vg_sdb\" not found\n Cannot process volume group gluster_vg_sdb\n", "index": 0, "item": {"lvname": "gluster_lv_engine", "size": "150G", "vgname": "gluster_vg_sdb"}, "msg": "Volume group gluster_vg_sdb does not exist.", "rc": 5} NO MORE HOSTS LEFT ************************************************************* NO MORE HOSTS LEFT ************************************************************* PLAY RECAP ********************************************************************* virtnodetest-0-0 : ok=19 changed=3 unreachable=0 failed=1 skipped=41 rescued=0 ignored=0 virtnodetest-0-1 : ok=20 changed=4 unreachable=0 failed=1 skipped=41 rescued=0 ignored=0 virtnodetest-0-2 : ok=19 changed=3 unreachable=0 failed=1 skipped=41 rescued=0 ignored=0 Please check /var/log/cockpit/ovirt-dashboard/gluster-deployment.log for more informations. ====================================================================== How can we resolve this issue?