4.3.10 is the latest version of oVirt 4.3, not sure why you'd want to try installing an older version.

These types of installer errors are usually caused by either the device 'sdb' is smaller than you're trying to allocate
or
has some previous partition formatting on it (installer presumes a non-formatted drive, and stops to protect data)
or
the system has lvm or multipathd filtering out the drive /dev/sdb
In the second case usually wipefs can be used to make /dev/sdb clean for script to work.
for the third case, you may need to edit the /etc/multipathd.conf or lvm.conf and reboot to make
the device available to the installer.



On Thu, May 13, 2021 at 7:23 AM techbreak--- via Users <users@ovirt.org> wrote:
As title, we want to use 3 hosts with hyperconverged solution of oVirt. We installed oVirt and Gluster as reported in the guide https://www.ovirt.org/documentation/gluster-hyperconverged/chap-Deploying_Hyperconverged.html
When we try to deploy, we get some errors which we cannot figure out.

==============================================================
gdeploy creates as configuration rules:
hc_nodes:
  hosts:
    virtnodetest-0-0:
      gluster_infra_volume_groups:
        - vgname: gluster_vg_sdb
          pvname: /dev/sdb
      gluster_infra_mount_devices:
        - path: /gluster_bricks/engine
          lvname: gluster_lv_engine
          vgname: gluster_vg_sdb
        - path: /gluster_bricks/isostorage
          lvname: gluster_lv_isostorage
          vgname: gluster_vg_sdb
        - path: /gluster_bricks/vmstorage
          lvname: gluster_lv_vmstorage
          vgname: gluster_vg_sdb
      gluster_infra_thick_lvs:
        - vgname: gluster_vg_sdb
          lvname: gluster_lv_engine
          size: 150G
      gluster_infra_thinpools:
        - vgname: gluster_vg_sdb
          thinpoolname: gluster_thinpool_gluster_vg_sdb
          poolmetadatasize: 16G
      gluster_infra_lv_logicalvols:
        - vgname: gluster_vg_sdb
          thinpool: gluster_thinpool_gluster_vg_sdb
          lvname: gluster_lv_isostorage
          lvsize: 250G
        - vgname: gluster_vg_sdb
          thinpool: gluster_thinpool_gluster_vg_sdb
          lvname: gluster_lv_vmstorage
          lvsize: 3500G
    virtnodetest-0-1:
      gluster_infra_volume_groups:
        - vgname: gluster_vg_sdb
          pvname: /dev/sdb
      gluster_infra_mount_devices:
        - path: /gluster_bricks/engine
          lvname: gluster_lv_engine
          vgname: gluster_vg_sdb
        - path: /gluster_bricks/isostorage
          lvname: gluster_lv_isostorage
          vgname: gluster_vg_sdb
        - path: /gluster_bricks/vmstorage
          lvname: gluster_lv_vmstorage
          vgname: gluster_vg_sdb
      gluster_infra_thick_lvs:
        - vgname: gluster_vg_sdb
          lvname: gluster_lv_engine
          size: 150G
      gluster_infra_thinpools:
        - vgname: gluster_vg_sdb
          thinpoolname: gluster_thinpool_gluster_vg_sdb
          poolmetadatasize: 16G
      gluster_infra_lv_logicalvols:
        - vgname: gluster_vg_sdb
          thinpool: gluster_thinpool_gluster_vg_sdb
          lvname: gluster_lv_isostorage
          lvsize: 250G
        - vgname: gluster_vg_sdb
          thinpool: gluster_thinpool_gluster_vg_sdb
          lvname: gluster_lv_vmstorage
          lvsize: 3500G
    virtnodetest-0-2:
      gluster_infra_volume_groups:
        - vgname: gluster_vg_sdb
          pvname: /dev/sdb
      gluster_infra_mount_devices:
        - path: /gluster_bricks/engine
          lvname: gluster_lv_engine
          vgname: gluster_vg_sdb
        - path: /gluster_bricks/isostorage
          lvname: gluster_lv_isostorage
          vgname: gluster_vg_sdb
        - path: /gluster_bricks/vmstorage
          lvname: gluster_lv_vmstorage
          vgname: gluster_vg_sdb
      gluster_infra_thick_lvs:
        - vgname: gluster_vg_sdb
          lvname: gluster_lv_engine
          size: 150G
      gluster_infra_thinpools:
        - vgname: gluster_vg_sdb
          thinpoolname: gluster_thinpool_gluster_vg_sdb
          poolmetadatasize: 16G
      gluster_infra_lv_logicalvols:
        - vgname: gluster_vg_sdb
          thinpool: gluster_thinpool_gluster_vg_sdb
          lvname: gluster_lv_isostorage
          lvsize: 250G
        - vgname: gluster_vg_sdb
          thinpool: gluster_thinpool_gluster_vg_sdb
          lvname: gluster_lv_vmstorage
          lvsize: 3500G
  vars:
    gluster_infra_disktype: JBOD
    gluster_set_selinux_labels: true
    gluster_infra_fw_ports:
      - 2049/tcp
      - 54321/tcp
      - 5900/tcp
      - 5900-6923/tcp
      - 5666/tcp
      - 16514/tcp
    gluster_infra_fw_permanent: true
    gluster_infra_fw_state: enabled
    gluster_infra_fw_zone: public
    gluster_infra_fw_services:
      - glusterfs
    gluster_features_force_varlogsizecheck: false
    cluster_nodes:
      - virtnodetest-0-0
      - virtnodetest-0-1
      - virtnodetest-0-2
    gluster_features_hci_cluster: '{{ cluster_nodes }}'
    gluster_features_hci_volumes:
      - volname: engine
        brick: /gluster_bricks/engine/engine
        arbiter: 0
      - volname: isostorage
        brick: /gluster_bricks/isostorage/isostorage
        arbiter: 0
      - volname: vmstorage
        brick: /gluster_bricks/vmstorage/vmstorage
        arbiter: 0


=========================================================================
The system returns this error:

PLAY [Setup backend] ***********************************************************

TASK [Gathering Facts] *********************************************************
ok: [virtnodetest-0-2]
ok: [virtnodetest-0-1]
ok: [virtnodetest-0-0]

TASK [Check if valid hostnames are provided] ***********************************
changed: [virtnodetest-0-1] => (item=virtnodetest-0-1)
changed: [virtnodetest-0-1] => (item=virtnodetest-0-0)
changed: [virtnodetest-0-1] => (item=virtnodetest-0-2)

TASK [Check if provided hostnames are valid] ***********************************
ok: [virtnodetest-0-1] => {
    "changed": false,
    "msg": "All assertions passed"
}
ok: [virtnodetest-0-0] => {
    "changed": false,
    "msg": "All assertions passed"
}
ok: [virtnodetest-0-2] => {
    "changed": false,
    "msg": "All assertions passed"
}

TASK [Check if /var/log has enough disk space] *********************************
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]

TASK [Check if the /var is greater than 15G] ***********************************
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]

TASK [Check if disks have logical block size of 512B] **************************
skipping: [virtnodetest-0-1] => (item={u'vgname': u'gluster_vg_sdb', u'pvname': u'/dev/sdb'})
skipping: [virtnodetest-0-0] => (item={u'vgname': u'gluster_vg_sdb', u'pvname': u'/dev/sdb'})
skipping: [virtnodetest-0-2] => (item={u'vgname': u'gluster_vg_sdb', u'pvname': u'/dev/sdb'})

TASK [Check if logical block size is 512 bytes] ********************************
skipping: [virtnodetest-0-1] => (item=Logical Block Size)
skipping: [virtnodetest-0-0] => (item=Logical Block Size)
skipping: [virtnodetest-0-2] => (item=Logical Block Size)

TASK [Get logical block size of VDO devices] ***********************************
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]

TASK [Check if logical block size is 512 bytes for VDO devices] ****************
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]

TASK [gluster.infra/roles/firewall_config : Start firewalld if not already started] ***
ok: [virtnodetest-0-2]
ok: [virtnodetest-0-1]
ok: [virtnodetest-0-0]

TASK [gluster.infra/roles/firewall_config : check if required variables are set] ***
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]

TASK [gluster.infra/roles/firewall_config : Open/Close firewalld ports] ********
ok: [virtnodetest-0-2] => (item=2049/tcp)
ok: [virtnodetest-0-1] => (item=2049/tcp)
ok: [virtnodetest-0-0] => (item=2049/tcp)
ok: [virtnodetest-0-2] => (item=54321/tcp)
ok: [virtnodetest-0-0] => (item=54321/tcp)
ok: [virtnodetest-0-1] => (item=54321/tcp)
ok: [virtnodetest-0-2] => (item=5900/tcp)
ok: [virtnodetest-0-1] => (item=5900/tcp)
ok: [virtnodetest-0-0] => (item=5900/tcp)
ok: [virtnodetest-0-2] => (item=5900-6923/tcp)
ok: [virtnodetest-0-0] => (item=5900-6923/tcp)
ok: [virtnodetest-0-1] => (item=5900-6923/tcp)
ok: [virtnodetest-0-2] => (item=5666/tcp)
ok: [virtnodetest-0-1] => (item=5666/tcp)
ok: [virtnodetest-0-0] => (item=5666/tcp)
ok: [virtnodetest-0-2] => (item=16514/tcp)
ok: [virtnodetest-0-1] => (item=16514/tcp)
ok: [virtnodetest-0-0] => (item=16514/tcp)

TASK [gluster.infra/roles/firewall_config : Add/Delete services to firewalld rules] ***
ok: [virtnodetest-0-1] => (item=glusterfs)
ok: [virtnodetest-0-0] => (item=glusterfs)
ok: [virtnodetest-0-2] => (item=glusterfs)

TASK [gluster.infra/roles/backend_setup : Check if vdsm-python package is installed or not] ***
changed: [virtnodetest-0-2]
changed: [virtnodetest-0-1]
changed: [virtnodetest-0-0]

TASK [gluster.infra/roles/backend_setup : Remove the existing LVM filter] ******
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]

TASK [gluster.infra/roles/backend_setup : Check that the multipath.conf exists] ***
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]

TASK [gluster.infra/roles/backend_setup : Ensure that multipathd services is enabled if not] ***
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]

TASK [gluster.infra/roles/backend_setup : Ensure that multipathd services is running] ***
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]

TASK [gluster.infra/roles/backend_setup : Create /etc/multipath/conf.d if doesn't exists] ***
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]

TASK [gluster.infra/roles/backend_setup : Get the UUID of the devices] *********
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]

TASK [gluster.infra/roles/backend_setup : Check that the blacklist.conf exists] ***
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]

TASK [gluster.infra/roles/backend_setup : Create blacklist template content] ***
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]

TASK [gluster.infra/roles/backend_setup : Add wwid to blacklist in blacklist.conf file] ***
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]

TASK [gluster.infra/roles/backend_setup : Reload multipathd] *******************
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]

TASK [gluster.infra/roles/backend_setup : Gather facts to determine the OS distribution] ***
ok: [virtnodetest-0-1]
ok: [virtnodetest-0-0]
ok: [virtnodetest-0-2]

TASK [gluster.infra/roles/backend_setup : Change to Install lvm tools for debian systems.] ***
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]

TASK [gluster.infra/roles/backend_setup : Change to Install lvm tools for RHEL systems.] ***
ok: [virtnodetest-0-2]
ok: [virtnodetest-0-1]
ok: [virtnodetest-0-0]

TASK [gluster.infra/roles/backend_setup : Install python-yaml package for Debian systems] ***
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]

TASK [gluster.infra/roles/backend_setup : Initialize vdo_devs array] ***********
ok: [virtnodetest-0-1]
ok: [virtnodetest-0-0]
ok: [virtnodetest-0-2]

TASK [gluster.infra/roles/backend_setup : Record VDO devices (if any)] *********
skipping: [virtnodetest-0-1] => (item={u'vgname': u'gluster_vg_sdb', u'pvname': u'/dev/sdb'})
skipping: [virtnodetest-0-0] => (item={u'vgname': u'gluster_vg_sdb', u'pvname': u'/dev/sdb'})
skipping: [virtnodetest-0-2] => (item={u'vgname': u'gluster_vg_sdb', u'pvname': u'/dev/sdb'})

TASK [gluster.infra/roles/backend_setup : Configure lvm thinpool extend threshold] ***
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]

TASK [gluster.infra/roles/backend_setup : Configure lvm thinpool extend percentage] ***
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]

TASK [gluster.infra/roles/backend_setup : Check if vdo block device exists] ****
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]

TASK [gluster.infra/roles/backend_setup : Record for missing devices for phase 2] ***
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]

TASK [gluster.infra/roles/backend_setup : set fact if it will at least install 1 vdo device] ***

TASK [gluster.infra/roles/backend_setup : Install VDO dependencies] ************
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]

TASK [gluster.infra/roles/backend_setup : set fact about vdo installed deps] ***
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]

TASK [gluster.infra/roles/backend_setup : Enable and start vdo service] ********
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]

TASK [gluster.infra/roles/backend_setup : Create VDO with specified size] ******
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]

TASK [gluster.infra/roles/backend_setup : Set VDO maxDiscardSize as 16M] *******
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]

TASK [gluster.infra/roles/backend_setup : Stop VDO volumes] ********************
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]

TASK [gluster.infra/roles/backend_setup : Start VDO volumes] *******************
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]

TASK [gluster.infra/roles/backend_setup : Check if valid disktype is provided] ***
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]

TASK [gluster.infra/roles/backend_setup : Set PV data alignment for JBOD] ******
ok: [virtnodetest-0-1]
ok: [virtnodetest-0-0]
ok: [virtnodetest-0-2]

TASK [gluster.infra/roles/backend_setup : Set PV data alignment for RAID] ******
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]

TASK [gluster.infra/roles/backend_setup : Set VG physical extent size for RAID] ***
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]

TASK [gluster.infra/roles/backend_setup : include_tasks] ***********************
included: /etc/ansible/roles/gluster.infra/roles/backend_setup/tasks/get_vg_groupings.yml for virtnodetest-0-1, virtnodetest-0-0, virtnodetest-0-2

TASK [gluster.infra/roles/backend_setup : Group devices by volume group name, including existing devices] ***
ok: [virtnodetest-0-1]
ok: [virtnodetest-0-0]
ok: [virtnodetest-0-2]

TASK [gluster.infra/roles/backend_setup : Check if vg block device exists] *****
changed: [virtnodetest-0-0] => (item={u'key': u'gluster_vg_sdb', u'value': [{u'vgname': u'gluster_vg_sdb', u'pvname': u'/dev/sdb'}]})
changed: [virtnodetest-0-1] => (item={u'key': u'gluster_vg_sdb', u'value': [{u'vgname': u'gluster_vg_sdb', u'pvname': u'/dev/sdb'}]})
changed: [virtnodetest-0-2] => (item={u'key': u'gluster_vg_sdb', u'value': [{u'vgname': u'gluster_vg_sdb', u'pvname': u'/dev/sdb'}]})

TASK [gluster.infra/roles/backend_setup : Filter none-existing devices] ********
ok: [virtnodetest-0-1]
ok: [virtnodetest-0-0]
ok: [virtnodetest-0-2]

TASK [gluster.infra/roles/backend_setup : Record for missing devices for phase 2] ***
ok: [virtnodetest-0-1] => (item={u'stderr_lines': [], u'ansible_loop_var': u'item', u'end': u'2021-05-13 12:18:33.575598', u'stderr': u'', u'stdout': u'0', u'changed': True, u'failed': False, u'delta': u'0:00:00.009901', u'cmd': u' test -b /dev/sdb && echo "1" || echo  "0"; \n', u'item': {u'value': [{u'vgname': u'gluster_vg_sdb', u'pvname': u'/dev/sdb'}], u'key': u'gluster_vg_sdb'}, u'rc': 0, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u' test -b /dev/sdb && echo "1" || echo  "0"; \n', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'stdout_lines': [u'0'], u'start': u'2021-05-13 12:18:33.565697'})
ok: [virtnodetest-0-0] => (item={u'stderr_lines': [], u'ansible_loop_var': u'item', u'end': u'2021-05-13 10:52:56.886693', u'stderr': u'', u'stdout': u'0', u'changed': True, u'failed': False, u'delta': u'0:00:00.008123', u'cmd': u' test -b /dev/sdb && echo "1" || echo  "0"; \n', u'item': {u'value': [{u'vgname': u'gluster_vg_sdb', u'pvname': u'/dev/sdb'}], u'key': u'gluster_vg_sdb'}, u'rc': 0, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u' test -b /dev/sdb && echo "1" || echo  "0"; \n', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'stdout_lines': [u'0'], u'start': u'2021-05-13 10:52:56.878570'})
ok: [virtnodetest-0-2] => (item={u'stderr_lines': [], u'ansible_loop_var': u'item', u'end': u'2021-05-13 12:25:24.420710', u'stderr': u'', u'stdout': u'0', u'changed': True, u'failed': False, u'delta': u'0:00:00.007307', u'cmd': u' test -b /dev/sdb && echo "1" || echo  "0"; \n', u'item': {u'value': [{u'vgname': u'gluster_vg_sdb', u'pvname': u'/dev/sdb'}], u'key': u'gluster_vg_sdb'}, u'rc': 0, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u' test -b /dev/sdb && echo "1" || echo  "0"; \n', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'stdout_lines': [u'0'], u'start': u'2021-05-13 12:25:24.413403'})

TASK [gluster.infra/roles/backend_setup : Create volume groups] ****************
skipping: [virtnodetest-0-1] => (item={u'key': u'gluster_vg_sdb', u'value': []})
skipping: [virtnodetest-0-0] => (item={u'key': u'gluster_vg_sdb', u'value': []})
skipping: [virtnodetest-0-2] => (item={u'key': u'gluster_vg_sdb', u'value': []})

TASK [gluster.infra/roles/backend_setup : update LVM fact's] *******************
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]

TASK [gluster.infra/roles/backend_setup : Check if thick-lv block devices exists] ***
changed: [virtnodetest-0-0] => (item={u'lvname': u'gluster_lv_engine', u'vgname': u'gluster_vg_sdb', u'size': u'150G'})
changed: [virtnodetest-0-1] => (item={u'lvname': u'gluster_lv_engine', u'vgname': u'gluster_vg_sdb', u'size': u'150G'})
changed: [virtnodetest-0-2] => (item={u'lvname': u'gluster_lv_engine', u'vgname': u'gluster_vg_sdb', u'size': u'150G'})

TASK [gluster.infra/roles/backend_setup : Record for missing devices for phase 2] ***
skipping: [virtnodetest-0-1] => (item={u'stderr_lines': [], u'ansible_loop_var': u'item', u'end': u'2021-05-13 12:18:37.528159', u'stderr': u'', u'stdout': u'1', u'changed': True, u'failed': False, u'delta': u'0:00:00.010032', u'cmd': u' echo "1" \n', u'item': {u'lvname': u'gluster_lv_engine', u'vgname': u'gluster_vg_sdb', u'size': u'150G'}, u'rc': 0, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u' echo "1" \n', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'stdout_lines': [u'1'], u'start': u'2021-05-13 12:18:37.518127'})
skipping: [virtnodetest-0-0] => (item={u'stderr_lines': [], u'ansible_loop_var': u'item', u'end': u'2021-05-13 10:53:00.863436', u'stderr': u'', u'stdout': u'1', u'changed': True, u'failed': False, u'delta': u'0:00:00.007459', u'cmd': u' echo "1" \n', u'item': {u'lvname': u'gluster_lv_engine', u'vgname': u'gluster_vg_sdb', u'size': u'150G'}, u'rc': 0, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u' echo "1" \n', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'stdout_lines': [u'1'], u'start': u'2021-05-13 10:53:00.855977'})
skipping: [virtnodetest-0-2] => (item={u'stderr_lines': [], u'ansible_loop_var': u'item', u'end': u'2021-05-13 12:25:28.261106', u'stderr': u'', u'stdout': u'1', u'changed': True, u'failed': False, u'delta': u'0:00:00.007818', u'cmd': u' echo "1" \n', u'item': {u'lvname': u'gluster_lv_engine', u'vgname': u'gluster_vg_sdb', u'size': u'150G'}, u'rc': 0, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u' echo "1" \n', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'stdout_lines': [u'1'], u'start': u'2021-05-13 12:25:28.253288'})

TASK [gluster.infra/roles/backend_setup : include_tasks] ***********************
included: /etc/ansible/roles/gluster.infra/roles/backend_setup/tasks/get_vg_groupings.yml for virtnodetest-0-1, virtnodetest-0-0, virtnodetest-0-2

TASK [gluster.infra/roles/backend_setup : Group devices by volume group name, including existing devices] ***
ok: [virtnodetest-0-1]
ok: [virtnodetest-0-0]
ok: [virtnodetest-0-2]

TASK [gluster.infra/roles/backend_setup : Check if vg block device exists] *****

TASK [gluster.infra/roles/backend_setup : Filter none-existing devices] ********
ok: [virtnodetest-0-1]
ok: [virtnodetest-0-0]
ok: [virtnodetest-0-2]

TASK [gluster.infra/roles/backend_setup : Make sure thick pvs exists in volume group] ***

TASK [gluster.infra/roles/backend_setup : update LVM fact's] *******************
skipping: [virtnodetest-0-1]
skipping: [virtnodetest-0-0]
skipping: [virtnodetest-0-2]

TASK [gluster.infra/roles/backend_setup : Create thick logical volume] *********
failed: [virtnodetest-0-1] (item={u'lvname': u'gluster_lv_engine', u'vgname': u'gluster_vg_sdb', u'size': u'150G'}) => {"ansible_index_var": "index", "ansible_loop_var": "item", "changed": false, "err": "  WARNING: Device for PV gx6iUE-369Z-3FDP-aRUQ-Wur0-1Xhf-v4g79j not found or rejected by a filter.\n  Volume group \"gluster_vg_sdb\" not found\n  Cannot process volume group gluster_vg_sdb\n", "index": 0, "item": {"lvname": "gluster_lv_engine", "size": "150G", "vgname": "gluster_vg_sdb"}, "msg": "Volume group gluster_vg_sdb does not exist.", "rc": 5}
failed: [virtnodetest-0-0] (item={u'lvname': u'gluster_lv_engine', u'vgname': u'gluster_vg_sdb', u'size': u'150G'}) => {"ansible_index_var": "index", "ansible_loop_var": "item", "changed": false, "err": "  Volume group \"gluster_vg_sdb\" not found\n  Cannot process volume group gluster_vg_sdb\n", "index": 0, "item": {"lvname": "gluster_lv_engine", "size": "150G", "vgname": "gluster_vg_sdb"}, "msg": "Volume group gluster_vg_sdb does not exist.", "rc": 5}
failed: [virtnodetest-0-2] (item={u'lvname': u'gluster_lv_engine', u'vgname': u'gluster_vg_sdb', u'size': u'150G'}) => {"ansible_index_var": "index", "ansible_loop_var": "item", "changed": false, "err": "  Volume group \"gluster_vg_sdb\" not found\n  Cannot process volume group gluster_vg_sdb\n", "index": 0, "item": {"lvname": "gluster_lv_engine", "size": "150G", "vgname": "gluster_vg_sdb"}, "msg": "Volume group gluster_vg_sdb does not exist.", "rc": 5}

NO MORE HOSTS LEFT *************************************************************

NO MORE HOSTS LEFT *************************************************************

PLAY RECAP *********************************************************************
virtnodetest-0-0           : ok=19   changed=3    unreachable=0    failed=1    skipped=41   rescued=0    ignored=0   
virtnodetest-0-1           : ok=20   changed=4    unreachable=0    failed=1    skipped=41   rescued=0    ignored=0   
virtnodetest-0-2           : ok=19   changed=3    unreachable=0    failed=1    skipped=41   rescued=0    ignored=0   

Please check /var/log/cockpit/ovirt-dashboard/gluster-deployment.log for more informations.
======================================================================

How can we resolve this issue?
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-leave@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/
List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/3J3GA4SWOISUR57TDJIA7IZ2CLGCCFZR/