
4.3.10 is the latest version of oVirt 4.3, not sure why you'd want to try installing an older version. These types of installer errors are usually caused by either the device 'sdb' is smaller than you're trying to allocate or has some previous partition formatting on it (installer presumes a non-formatted drive, and stops to protect data) or the system has lvm or multipathd filtering out the drive /dev/sdb In the second case usually wipefs can be used to make /dev/sdb clean for script to work. for the third case, you may need to edit the /etc/multipathd.conf or lvm.conf and reboot to make the device available to the installer. On Thu, May 13, 2021 at 7:23 AM techbreak--- via Users <users@ovirt.org> wrote:
As title, we want to use 3 hosts with hyperconverged solution of oVirt. We installed oVirt and Gluster as reported in the guide https://www.ovirt.org/documentation/gluster-hyperconverged/chap-Deploying_Hy... When we try to deploy, we get some errors which we cannot figure out.
============================================================== gdeploy creates as configuration rules: hc_nodes: hosts: virtnodetest-0-0: gluster_infra_volume_groups: - vgname: gluster_vg_sdb pvname: /dev/sdb gluster_infra_mount_devices: - path: /gluster_bricks/engine lvname: gluster_lv_engine vgname: gluster_vg_sdb - path: /gluster_bricks/isostorage lvname: gluster_lv_isostorage vgname: gluster_vg_sdb - path: /gluster_bricks/vmstorage lvname: gluster_lv_vmstorage vgname: gluster_vg_sdb gluster_infra_thick_lvs: - vgname: gluster_vg_sdb lvname: gluster_lv_engine size: 150G gluster_infra_thinpools: - vgname: gluster_vg_sdb thinpoolname: gluster_thinpool_gluster_vg_sdb poolmetadatasize: 16G gluster_infra_lv_logicalvols: - vgname: gluster_vg_sdb thinpool: gluster_thinpool_gluster_vg_sdb lvname: gluster_lv_isostorage lvsize: 250G - vgname: gluster_vg_sdb thinpool: gluster_thinpool_gluster_vg_sdb lvname: gluster_lv_vmstorage lvsize: 3500G virtnodetest-0-1: gluster_infra_volume_groups: - vgname: gluster_vg_sdb pvname: /dev/sdb gluster_infra_mount_devices: - path: /gluster_bricks/engine lvname: gluster_lv_engine vgname: gluster_vg_sdb - path: /gluster_bricks/isostorage lvname: gluster_lv_isostorage vgname: gluster_vg_sdb - path: /gluster_bricks/vmstorage lvname: gluster_lv_vmstorage vgname: gluster_vg_sdb gluster_infra_thick_lvs: - vgname: gluster_vg_sdb lvname: gluster_lv_engine size: 150G gluster_infra_thinpools: - vgname: gluster_vg_sdb thinpoolname: gluster_thinpool_gluster_vg_sdb poolmetadatasize: 16G gluster_infra_lv_logicalvols: - vgname: gluster_vg_sdb thinpool: gluster_thinpool_gluster_vg_sdb lvname: gluster_lv_isostorage lvsize: 250G - vgname: gluster_vg_sdb thinpool: gluster_thinpool_gluster_vg_sdb lvname: gluster_lv_vmstorage lvsize: 3500G virtnodetest-0-2: gluster_infra_volume_groups: - vgname: gluster_vg_sdb pvname: /dev/sdb gluster_infra_mount_devices: - path: /gluster_bricks/engine lvname: gluster_lv_engine vgname: gluster_vg_sdb - path: /gluster_bricks/isostorage lvname: gluster_lv_isostorage vgname: gluster_vg_sdb - path: /gluster_bricks/vmstorage lvname: gluster_lv_vmstorage vgname: gluster_vg_sdb gluster_infra_thick_lvs: - vgname: gluster_vg_sdb lvname: gluster_lv_engine size: 150G gluster_infra_thinpools: - vgname: gluster_vg_sdb thinpoolname: gluster_thinpool_gluster_vg_sdb poolmetadatasize: 16G gluster_infra_lv_logicalvols: - vgname: gluster_vg_sdb thinpool: gluster_thinpool_gluster_vg_sdb lvname: gluster_lv_isostorage lvsize: 250G - vgname: gluster_vg_sdb thinpool: gluster_thinpool_gluster_vg_sdb lvname: gluster_lv_vmstorage lvsize: 3500G vars: gluster_infra_disktype: JBOD gluster_set_selinux_labels: true gluster_infra_fw_ports: - 2049/tcp - 54321/tcp - 5900/tcp - 5900-6923/tcp - 5666/tcp - 16514/tcp gluster_infra_fw_permanent: true gluster_infra_fw_state: enabled gluster_infra_fw_zone: public gluster_infra_fw_services: - glusterfs gluster_features_force_varlogsizecheck: false cluster_nodes: - virtnodetest-0-0 - virtnodetest-0-1 - virtnodetest-0-2 gluster_features_hci_cluster: '{{ cluster_nodes }}' gluster_features_hci_volumes: - volname: engine brick: /gluster_bricks/engine/engine arbiter: 0 - volname: isostorage brick: /gluster_bricks/isostorage/isostorage arbiter: 0 - volname: vmstorage brick: /gluster_bricks/vmstorage/vmstorage arbiter: 0
========================================================================= The system returns this error:
PLAY [Setup backend] ***********************************************************
TASK [Gathering Facts] ********************************************************* ok: [virtnodetest-0-2] ok: [virtnodetest-0-1] ok: [virtnodetest-0-0]
TASK [Check if valid hostnames are provided] *********************************** changed: [virtnodetest-0-1] => (item=virtnodetest-0-1) changed: [virtnodetest-0-1] => (item=virtnodetest-0-0) changed: [virtnodetest-0-1] => (item=virtnodetest-0-2)
TASK [Check if provided hostnames are valid] *********************************** ok: [virtnodetest-0-1] => { "changed": false, "msg": "All assertions passed" } ok: [virtnodetest-0-0] => { "changed": false, "msg": "All assertions passed" } ok: [virtnodetest-0-2] => { "changed": false, "msg": "All assertions passed" }
TASK [Check if /var/log has enough disk space] ********************************* skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2]
TASK [Check if the /var is greater than 15G] *********************************** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2]
TASK [Check if disks have logical block size of 512B] ************************** skipping: [virtnodetest-0-1] => (item={u'vgname': u'gluster_vg_sdb', u'pvname': u'/dev/sdb'}) skipping: [virtnodetest-0-0] => (item={u'vgname': u'gluster_vg_sdb', u'pvname': u'/dev/sdb'}) skipping: [virtnodetest-0-2] => (item={u'vgname': u'gluster_vg_sdb', u'pvname': u'/dev/sdb'})
TASK [Check if logical block size is 512 bytes] ******************************** skipping: [virtnodetest-0-1] => (item=Logical Block Size) skipping: [virtnodetest-0-0] => (item=Logical Block Size) skipping: [virtnodetest-0-2] => (item=Logical Block Size)
TASK [Get logical block size of VDO devices] *********************************** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2]
TASK [Check if logical block size is 512 bytes for VDO devices] **************** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/firewall_config : Start firewalld if not already started] *** ok: [virtnodetest-0-2] ok: [virtnodetest-0-1] ok: [virtnodetest-0-0]
TASK [gluster.infra/roles/firewall_config : check if required variables are set] *** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/firewall_config : Open/Close firewalld ports] ******** ok: [virtnodetest-0-2] => (item=2049/tcp) ok: [virtnodetest-0-1] => (item=2049/tcp) ok: [virtnodetest-0-0] => (item=2049/tcp) ok: [virtnodetest-0-2] => (item=54321/tcp) ok: [virtnodetest-0-0] => (item=54321/tcp) ok: [virtnodetest-0-1] => (item=54321/tcp) ok: [virtnodetest-0-2] => (item=5900/tcp) ok: [virtnodetest-0-1] => (item=5900/tcp) ok: [virtnodetest-0-0] => (item=5900/tcp) ok: [virtnodetest-0-2] => (item=5900-6923/tcp) ok: [virtnodetest-0-0] => (item=5900-6923/tcp) ok: [virtnodetest-0-1] => (item=5900-6923/tcp) ok: [virtnodetest-0-2] => (item=5666/tcp) ok: [virtnodetest-0-1] => (item=5666/tcp) ok: [virtnodetest-0-0] => (item=5666/tcp) ok: [virtnodetest-0-2] => (item=16514/tcp) ok: [virtnodetest-0-1] => (item=16514/tcp) ok: [virtnodetest-0-0] => (item=16514/tcp)
TASK [gluster.infra/roles/firewall_config : Add/Delete services to firewalld rules] *** ok: [virtnodetest-0-1] => (item=glusterfs) ok: [virtnodetest-0-0] => (item=glusterfs) ok: [virtnodetest-0-2] => (item=glusterfs)
TASK [gluster.infra/roles/backend_setup : Check if vdsm-python package is installed or not] *** changed: [virtnodetest-0-2] changed: [virtnodetest-0-1] changed: [virtnodetest-0-0]
TASK [gluster.infra/roles/backend_setup : Remove the existing LVM filter] ****** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Check that the multipath.conf exists] *** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Ensure that multipathd services is enabled if not] *** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Ensure that multipathd services is running] *** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Create /etc/multipath/conf.d if doesn't exists] *** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Get the UUID of the devices] ********* skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Check that the blacklist.conf exists] *** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Create blacklist template content] *** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Add wwid to blacklist in blacklist.conf file] *** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Reload multipathd] ******************* skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Gather facts to determine the OS distribution] *** ok: [virtnodetest-0-1] ok: [virtnodetest-0-0] ok: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Change to Install lvm tools for debian systems.] *** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Change to Install lvm tools for RHEL systems.] *** ok: [virtnodetest-0-2] ok: [virtnodetest-0-1] ok: [virtnodetest-0-0]
TASK [gluster.infra/roles/backend_setup : Install python-yaml package for Debian systems] *** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Initialize vdo_devs array] *********** ok: [virtnodetest-0-1] ok: [virtnodetest-0-0] ok: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Record VDO devices (if any)] ********* skipping: [virtnodetest-0-1] => (item={u'vgname': u'gluster_vg_sdb', u'pvname': u'/dev/sdb'}) skipping: [virtnodetest-0-0] => (item={u'vgname': u'gluster_vg_sdb', u'pvname': u'/dev/sdb'}) skipping: [virtnodetest-0-2] => (item={u'vgname': u'gluster_vg_sdb', u'pvname': u'/dev/sdb'})
TASK [gluster.infra/roles/backend_setup : Configure lvm thinpool extend threshold] *** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Configure lvm thinpool extend percentage] *** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Check if vdo block device exists] **** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Record for missing devices for phase 2] *** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : set fact if it will at least install 1 vdo device] ***
TASK [gluster.infra/roles/backend_setup : Install VDO dependencies] ************ skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : set fact about vdo installed deps] *** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Enable and start vdo service] ******** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Create VDO with specified size] ****** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Set VDO maxDiscardSize as 16M] ******* skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Stop VDO volumes] ******************** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Start VDO volumes] ******************* skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Check if valid disktype is provided] *** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Set PV data alignment for JBOD] ****** ok: [virtnodetest-0-1] ok: [virtnodetest-0-0] ok: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Set PV data alignment for RAID] ****** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Set VG physical extent size for RAID] *** skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : include_tasks] *********************** included: /etc/ansible/roles/gluster.infra/roles/backend_setup/tasks/get_vg_groupings.yml for virtnodetest-0-1, virtnodetest-0-0, virtnodetest-0-2
TASK [gluster.infra/roles/backend_setup : Group devices by volume group name, including existing devices] *** ok: [virtnodetest-0-1] ok: [virtnodetest-0-0] ok: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Check if vg block device exists] ***** changed: [virtnodetest-0-0] => (item={u'key': u'gluster_vg_sdb', u'value': [{u'vgname': u'gluster_vg_sdb', u'pvname': u'/dev/sdb'}]}) changed: [virtnodetest-0-1] => (item={u'key': u'gluster_vg_sdb', u'value': [{u'vgname': u'gluster_vg_sdb', u'pvname': u'/dev/sdb'}]}) changed: [virtnodetest-0-2] => (item={u'key': u'gluster_vg_sdb', u'value': [{u'vgname': u'gluster_vg_sdb', u'pvname': u'/dev/sdb'}]})
TASK [gluster.infra/roles/backend_setup : Filter none-existing devices] ******** ok: [virtnodetest-0-1] ok: [virtnodetest-0-0] ok: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Record for missing devices for phase 2] *** ok: [virtnodetest-0-1] => (item={u'stderr_lines': [], u'ansible_loop_var': u'item', u'end': u'2021-05-13 12:18:33.575598', u'stderr': u'', u'stdout': u'0', u'changed': True, u'failed': False, u'delta': u'0:00:00.009901', u'cmd': u' test -b /dev/sdb && echo "1" || echo "0"; \n', u'item': {u'value': [{u'vgname': u'gluster_vg_sdb', u'pvname': u'/dev/sdb'}], u'key': u'gluster_vg_sdb'}, u'rc': 0, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u' test -b /dev/sdb && echo "1" || echo "0"; \n', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'stdout_lines': [u'0'], u'start': u'2021-05-13 12:18:33.565697'}) ok: [virtnodetest-0-0] => (item={u'stderr_lines': [], u'ansible_loop_var': u'item', u'end': u'2021-05-13 10:52:56.886693', u'stderr': u'', u'stdout': u'0', u'changed': True, u'failed': False, u'delta': u'0:00:00.008123', u'cmd': u' test -b /dev/sdb && echo "1" || echo "0"; \n', u'item': {u'value': [{u'vgname': u'gluster_vg_sdb', u'pvname': u'/dev/sdb'}], u'key': u'gluster_vg_sdb'}, u'rc': 0, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u' test -b /dev/sdb && echo "1" || echo "0"; \n', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'stdout_lines': [u'0'], u'start': u'2021-05-13 10:52:56.878570'}) ok: [virtnodetest-0-2] => (item={u'stderr_lines': [], u'ansible_loop_var': u'item', u'end': u'2021-05-13 12:25:24.420710', u'stderr': u'', u'stdout': u'0', u'changed': True, u'failed': False, u'delta': u'0:00:00.007307', u'cmd': u' test -b /dev/sdb && echo "1" || echo "0"; \n', u'item': {u'value': [{u'vgname': u'gluster_vg_sdb', u'pvname': u'/dev/sdb'}], u'key': u'gluster_vg_sdb'}, u'rc': 0, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u' test -b /dev/sdb && echo "1" || echo "0"; \n', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'stdout_lines': [u'0'], u'start': u'2021-05-13 12:25:24.413403'})
TASK [gluster.infra/roles/backend_setup : Create volume groups] **************** skipping: [virtnodetest-0-1] => (item={u'key': u'gluster_vg_sdb', u'value': []}) skipping: [virtnodetest-0-0] => (item={u'key': u'gluster_vg_sdb', u'value': []}) skipping: [virtnodetest-0-2] => (item={u'key': u'gluster_vg_sdb', u'value': []})
TASK [gluster.infra/roles/backend_setup : update LVM fact's] ******************* skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Check if thick-lv block devices exists] *** changed: [virtnodetest-0-0] => (item={u'lvname': u'gluster_lv_engine', u'vgname': u'gluster_vg_sdb', u'size': u'150G'}) changed: [virtnodetest-0-1] => (item={u'lvname': u'gluster_lv_engine', u'vgname': u'gluster_vg_sdb', u'size': u'150G'}) changed: [virtnodetest-0-2] => (item={u'lvname': u'gluster_lv_engine', u'vgname': u'gluster_vg_sdb', u'size': u'150G'})
TASK [gluster.infra/roles/backend_setup : Record for missing devices for phase 2] *** skipping: [virtnodetest-0-1] => (item={u'stderr_lines': [], u'ansible_loop_var': u'item', u'end': u'2021-05-13 12:18:37.528159', u'stderr': u'', u'stdout': u'1', u'changed': True, u'failed': False, u'delta': u'0:00:00.010032', u'cmd': u' echo "1" \n', u'item': {u'lvname': u'gluster_lv_engine', u'vgname': u'gluster_vg_sdb', u'size': u'150G'}, u'rc': 0, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u' echo "1" \n', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'stdout_lines': [u'1'], u'start': u'2021-05-13 12:18:37.518127'}) skipping: [virtnodetest-0-0] => (item={u'stderr_lines': [], u'ansible_loop_var': u'item', u'end': u'2021-05-13 10:53:00.863436', u'stderr': u'', u'stdout': u'1', u'changed': True, u'failed': False, u'delta': u'0:00:00.007459', u'cmd': u' echo "1" \n', u'item': {u'lvname': u'gluster_lv_engine', u'vgname': u'gluster_vg_sdb', u'size': u'150G'}, u'rc': 0, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u' echo "1" \n', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'stdout_lines': [u'1'], u'start': u'2021-05-13 10:53:00.855977'}) skipping: [virtnodetest-0-2] => (item={u'stderr_lines': [], u'ansible_loop_var': u'item', u'end': u'2021-05-13 12:25:28.261106', u'stderr': u'', u'stdout': u'1', u'changed': True, u'failed': False, u'delta': u'0:00:00.007818', u'cmd': u' echo "1" \n', u'item': {u'lvname': u'gluster_lv_engine', u'vgname': u'gluster_vg_sdb', u'size': u'150G'}, u'rc': 0, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u' echo "1" \n', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'stdout_lines': [u'1'], u'start': u'2021-05-13 12:25:28.253288'})
TASK [gluster.infra/roles/backend_setup : include_tasks] *********************** included: /etc/ansible/roles/gluster.infra/roles/backend_setup/tasks/get_vg_groupings.yml for virtnodetest-0-1, virtnodetest-0-0, virtnodetest-0-2
TASK [gluster.infra/roles/backend_setup : Group devices by volume group name, including existing devices] *** ok: [virtnodetest-0-1] ok: [virtnodetest-0-0] ok: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Check if vg block device exists] *****
TASK [gluster.infra/roles/backend_setup : Filter none-existing devices] ******** ok: [virtnodetest-0-1] ok: [virtnodetest-0-0] ok: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Make sure thick pvs exists in volume group] ***
TASK [gluster.infra/roles/backend_setup : update LVM fact's] ******************* skipping: [virtnodetest-0-1] skipping: [virtnodetest-0-0] skipping: [virtnodetest-0-2]
TASK [gluster.infra/roles/backend_setup : Create thick logical volume] ********* failed: [virtnodetest-0-1] (item={u'lvname': u'gluster_lv_engine', u'vgname': u'gluster_vg_sdb', u'size': u'150G'}) => {"ansible_index_var": "index", "ansible_loop_var": "item", "changed": false, "err": " WARNING: Device for PV gx6iUE-369Z-3FDP-aRUQ-Wur0-1Xhf-v4g79j not found or rejected by a filter.\n Volume group \"gluster_vg_sdb\" not found\n Cannot process volume group gluster_vg_sdb\n", "index": 0, "item": {"lvname": "gluster_lv_engine", "size": "150G", "vgname": "gluster_vg_sdb"}, "msg": "Volume group gluster_vg_sdb does not exist.", "rc": 5} failed: [virtnodetest-0-0] (item={u'lvname': u'gluster_lv_engine', u'vgname': u'gluster_vg_sdb', u'size': u'150G'}) => {"ansible_index_var": "index", "ansible_loop_var": "item", "changed": false, "err": " Volume group \"gluster_vg_sdb\" not found\n Cannot process volume group gluster_vg_sdb\n", "index": 0, "item": {"lvname": "gluster_lv_engine", "size": "150G", "vgname": "gluster_vg_sdb"}, "msg": "Volume group gluster_vg_sdb does not exist.", "rc": 5} failed: [virtnodetest-0-2] (item={u'lvname': u'gluster_lv_engine', u'vgname': u'gluster_vg_sdb', u'size': u'150G'}) => {"ansible_index_var": "index", "ansible_loop_var": "item", "changed": false, "err": " Volume group \"gluster_vg_sdb\" not found\n Cannot process volume group gluster_vg_sdb\n", "index": 0, "item": {"lvname": "gluster_lv_engine", "size": "150G", "vgname": "gluster_vg_sdb"}, "msg": "Volume group gluster_vg_sdb does not exist.", "rc": 5}
NO MORE HOSTS LEFT *************************************************************
NO MORE HOSTS LEFT *************************************************************
PLAY RECAP ********************************************************************* virtnodetest-0-0 : ok=19 changed=3 unreachable=0 failed=1 skipped=41 rescued=0 ignored=0 virtnodetest-0-1 : ok=20 changed=4 unreachable=0 failed=1 skipped=41 rescued=0 ignored=0 virtnodetest-0-2 : ok=19 changed=3 unreachable=0 failed=1 skipped=41 rescued=0 ignored=0
Please check /var/log/cockpit/ovirt-dashboard/gluster-deployment.log for more informations. ======================================================================
How can we resolve this issue? _______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-leave@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/3J3GA4SWOISUR5...